From c492ed353ddcd974357e2c12682f147db1c8d873 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 14:08:36 -0600 Subject: [PATCH 01/69] Paper meta data correction for 2023.iwcs-1.29, closes #2867. --- data/xml/2023.iwcs.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.iwcs.xml b/data/xml/2023.iwcs.xml index cbe88ea29c..6e912e842a 100644 --- a/data/xml/2023.iwcs.xml +++ b/data/xml/2023.iwcs.xml @@ -288,7 +288,7 @@ opitz-etal-2023-smaragd - <fixed-case>AMR</fixed-case>4<fixed-case>NLI</fixed-case>: Interpretable and robust <fixed-case>NLI</fixed-case> measures from semantic graph + <fixed-case>AMR</fixed-case>4<fixed-case>NLI</fixed-case>: Interpretable and robust <fixed-case>NLI</fixed-case> measures from semantic graphs JuriOpitz ShiraWein JuliusSteen From a0abe8884ffa64d8880e67c2f1fa451c777b9112 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 14:16:41 -0600 Subject: [PATCH 02/69] Paper pdf correction for 2023.inlg-main.21, closes #2871. --- data/xml/2023.inlg.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.inlg.xml b/data/xml/2023.inlg.xml index 047d41f619..d5bdf65766 100644 --- a/data/xml/2023.inlg.xml +++ b/data/xml/2023.inlg.xml @@ -274,9 +274,11 @@ AlbertGatt 293–312 Current captioning datasets focus on object-centric captions, describing the visible objects in the image, often ending up stating the obvious (for humans), e.g. “people eating food in a park”. Although these datasets are useful to evaluate the ability of Vision & Language models to recognize and describe visual content, they do not support controlled experiments involving model testing or fine-tuning, with more high-level captions, which humans find easy and natural to produce. For example, people often describe images based on the type of scene they depict (“people at a holiday resort”) and the actions they perform (“people having a picnic”). Such concepts are based on personal experience and contribute to forming common sense assumptions. We present the High-Level Dataset, a dataset extending 14997 images from the COCO dataset, aligned with a new set of 134,973 human-annotated (high-level) captions collected along three axes: scenes, actions and rationales. We further extend this dataset with confidence scores collected from an independent set of readers, as well as a set of narrative captions generated synthetically, by combining each of the three axes. We describe this dataset and analyse it extensively. We also present baseline results for the High-Level Captioning task. - 2023.inlg-main.21 + 2023.inlg-main.21 cafagna-etal-2023-hl 10.18653/v1/2023.inlg-main.21 + + Updated Acknowledgments. Validating Predictive Models Of Evaluative Language For Controllable <fixed-case>D</fixed-case>ata2<fixed-case>T</fixed-case>ext Generation From a8560973d8e0a9ffed710da8b59a138a55d4aeb4 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 14:19:32 -0600 Subject: [PATCH 03/69] Paper meta data correction for 2023.ldk-1.46, closes #2877. --- data/xml/2023.ldk.xml | 1 + 1 file changed, 1 insertion(+) diff --git a/data/xml/2023.ldk.xml b/data/xml/2023.ldk.xml index a9b523cc09..0ba1d971b3 100644 --- a/data/xml/2023.ldk.xml +++ b/data/xml/2023.ldk.xml @@ -451,6 +451,7 @@ PurificaçãoSilvano DimitarTrajanov Ciprian-OctavianTruica + Elena-SimonaApostol ChristianChiarcos AnnaBaczkowska 434-439 From 800fc180a7a79a44ecc3f2e8833af73ec095b826 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 15:55:58 -0600 Subject: [PATCH 04/69] Paper meta data correction for P17-1159, closes #2881. --- data/xml/P17.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/P17.xml b/data/xml/P17.xml index 7a59ee9b5a..cb3947039b 100644 --- a/data/xml/P17.xml +++ b/data/xml/P17.xml @@ -2213,7 +2213,7 @@ two word-vectors results in a vector that is only a small angle away from the ve Cora - <fixed-case>U</fixed-case>niversal <fixed-case>D</fixed-case>ependencies Parsing for Colloquial Singaporean <fixed-case>E</fixed-case>nglish + <fixed-case>U</fixed-case>niversal <fixed-case>D</fixed-case>ependencies Parsing for Colloquial <fixed-case>S</fixed-case>ingaporean <fixed-case>E</fixed-case>nglish HongminWang YueZhang GuangYong LeonardChan From 64eb10eeaf5047331212c7a2f201cc5d22876c2d Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 16:01:22 -0600 Subject: [PATCH 05/69] Paper meta data and pdf correction for 2023.insights-1.1, closes #2883. --- data/xml/2023.insights.xml | 43 +++++++++++++++++++++++++++++++++++++- 1 file changed, 42 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.insights.xml b/data/xml/2023.insights.xml index 5934a7bdf1..7227af97b4 100644 --- a/data/xml/2023.insights.xml +++ b/data/xml/2023.insights.xml @@ -25,12 +25,53 @@ AnyaBelzADAPT Research Centre, Dublin City University CraigThomsonUniversity of Aberdeen EhudReiterUniversity of Aberdeen + GavinAbercrombieHeriot-Watt University + Jose M.Alonso-MoralUniversidade de Santiago de Compostela + MohammadArvanUniversity of Illinois Chicago + AnouckBraggaarTilburg University + MarkCieliebakZurich University of Applied Sciences + ElizabethClarkGoogle Research + Keesvan DeemterUtrecht University + TanviDinkarHeriot-Watt University + OndřejDušekCharles University Prague + SteffenEgerBielefeld University + QixiangFangUtrecht University + MingqiGaoPeking University + AlbertGattUtrecht University + DimitraGkatziaEdinburgh Napier University + JavierGonzález-CorbelleUniversidade de Santiago de Compostela + DirkHovyBocconi University + ManuelaHürlimannZurich University of Applied Sciences + TakumiItoTohoku University + John D.KelleherTechnological University Dublin + FilipKlubickaTechnological University Dublin + EmielKrahmerTilburg University + HuiyuanLaiGroningen University + Chrisvan der LeeTilburg University + YiruLiGroningen University + SaadMahamoodtrivago + MargotMieskesUniversity of Applied Sciences Darmstadt + Emielvan MiltenburgTilburg University + PabloMosteiroUtrecht University + MalvinaNissimGroningen University + NataliePardeUniversity of Illinois Chicago + OndřejPlátekCharles University Prague + VerenaRieserHeriot-Watt University + JieRuanPeking University + JoelTetreaultDataminr + AntonioToralGroningen University + XiaojunWanPeking University + LeoWannerUniversitat Pompeu Fabra + LewisWatsonEdinburgh Napier University + DiyiYangGeorgia Tech 1-10 We report our efforts in identifying a set of previous human evaluations in NLP that would be suitable for a coordinated study examining what makes human evaluations in NLP more/less reproducible. We present our results and findings, which include that just 13% of papers had (i) sufficiently low barriers to reproduction, and (ii) enough obtainable information, to be considered for reproduction, and that all but one of the experiments we selected for reproduction was discovered to have flaws that made the meaningfulness of conducting a reproduction questionable. As a result, we had to change our coordinated study design from a reproduce approach to a standardise-then-reproduce-twice approach. Our overall (negative) finding that the great majority of human evaluations in NLP is not repeatable and/or not reproducible and/or too flawed to justify reproduction, paints a dire picture, but presents an opportunity for a rethink about how to design and report human evaluations in NLP. - 2023.insights-1.1 + 2023.insights-1.1 belz-etal-2023-missing <fixed-case>ERATE</fixed-case>: Efficient Retrieval Augmented Text Embeddings From fd3b6d4b12823ef444a84b365b668774d62c1115 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 16:02:55 -0600 Subject: [PATCH 06/69] Paper meta data correction for 2023.semeval-1.179, closes #2885. --- data/xml/2023.semeval.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.semeval.xml b/data/xml/2023.semeval.xml index cb52317cd2..cecd51d7ef 100644 --- a/data/xml/2023.semeval.xml +++ b/data/xml/2023.semeval.xml @@ -2065,7 +2065,7 @@ <fixed-case>MDC</fixed-case> at <fixed-case>S</fixed-case>em<fixed-case>E</fixed-case>val-2023 Task 7: Fine-tuning Transformers for Textual Entailment Prediction and Evidence Retrieval in Clinical Trials RobertBevanMedicines Discovery Catapult - OisnTurbittMedicines Discovery Catapult + OisínTurbittMedicines Discovery Catapult MouhamadAboshokorMedicines Discovery Catapult 1287-1292 We present our entry to the Multi-evidence Natural Language Inference for Clinical Trial Datatask at SemEval 2023. We submitted entries forboth the evidence retrieval and textual entailment sub-tasks. For the evidence retrieval task,we fine-tuned the PubMedBERT transformermodel to extract relevant evidence from clinicaltrial data given a hypothesis concerning either asingle clinical trial or pair of clinical trials. Ourbest performing model achieved an F1 scoreof 0.804. For the textual entailment task, inwhich systems had to predict whether a hypothesis about either a single clinical trial or pair ofclinical trials is true or false, we fine-tuned theBioLinkBERT transformer model. We passedour evidence retrieval model’s output into ourtextual entailment model and submitted its output for the evaluation. Our best performingmodel achieved an F1 score of 0.695. From 9f9bb144b294f6f03ae8d4a6ebad965e327ef97e Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 19:26:30 -0600 Subject: [PATCH 07/69] Paper meta data correction for 2023.sigdial-1.27, closes #2895. --- data/xml/2023.sigdial.xml | 1 + 1 file changed, 1 insertion(+) diff --git a/data/xml/2023.sigdial.xml b/data/xml/2023.sigdial.xml index cb81fe1c3e..ad0c335b1b 100644 --- a/data/xml/2023.sigdial.xml +++ b/data/xml/2023.sigdial.xml @@ -333,6 +333,7 @@ Syndicom: Improving Conversational Commonsense with Error-Injection and Natural Language Feedback ChristopherRichardson + AnirudhSundar LarryHeck 297–308 Commonsense reasoning is a critical aspect of human communication. Despite recent advances in conversational AI driven by large language models, commonsense reasoning remains a challenging task. In this work, we introduce Syndicom - a method for improving commonsense in dialogue response generation. Syndicom consists of two components. The first component is a dataset composed of commonsense dialogues created from a knowledge graph and synthesized into natural language. This dataset includes both valid and invalid responses to dialogue contexts, along with natural language feedback (NLF) for the invalid responses. The second contribution is a two-step procedure: training a model to predict natural language feedback (NLF) for invalid responses, and then training a response generation model conditioned on the predicted NLF, the invalid response, and the dialogue. Syndicom is scalable and does not require reinforcement learning. Empirical results on three tasks are evaluated using a broad range of metrics. Syndicom achieves a relative improvement of 53% over ChatGPT on ROUGE-1, and human evaluators prefer Syndicom over ChatGPT 57% of the time. We will publicly release the code and the full dataset. From c0bce3c52397b23d05eff3373fbcabfb0bb4759e Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 19:29:26 -0600 Subject: [PATCH 08/69] Paper meta data correction for 2023.conll-1.25, closes #2897. --- data/xml/2023.conll.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.conll.xml b/data/xml/2023.conll.xml index 42034e72c4..95f9050dd0 100644 --- a/data/xml/2023.conll.xml +++ b/data/xml/2023.conll.xml @@ -274,7 +274,7 @@ TomKouwenhoven Wernerde Valk MarcoSpruit - PetervanderPutten + Petervan der Putten 389–402 To what degree should we ascribe cognitive capacities to Large Language Models (LLMs), such as the ability to reason about intentions and beliefs known as Theory of Mind (ToM)? Here we add to this emerging debate by (i) testing 11 base- and instruction-tuned LLMs on capabilities relevant to ToM beyond the dominant false-belief paradigm, including non-literal language usage and recursive intentionality; (ii) using newly rewritten versions of standardized tests to gauge LLMs’ robustness; (iii) prompting and scoring for open besides closed questions; and (iv) benchmarking LLM performance against that of children aged 7-10 on the same tasks. We find that instruction-tuned LLMs from the GPT family outperform other models, and often also children. Base-LLMs are mostly unable to solve ToM tasks, even with specialized prompting. We suggest that the interlinked evolution and development of language and ToM may help explain what instruction-tuning adds: rewarding cooperative communication that takes into account interlocutor and context. We conclude by arguing for a nuanced perspective on ToM in LLMs. 2023.conll-1.25 From 46d72a2e270e578bb864f59a8c8760c51afbca4c Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 19:33:49 -0600 Subject: [PATCH 09/69] Paper meta data correction for 2022.emnlp-main.744, closes #2899. --- data/xml/2022.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2022.emnlp.xml b/data/xml/2022.emnlp.xml index 4faab06072..63ef3add94 100644 --- a/data/xml/2022.emnlp.xml +++ b/data/xml/2022.emnlp.xml @@ -9701,7 +9701,7 @@ <fixed-case>KOLD</fixed-case>: <fixed-case>K</fixed-case>orean Offensive Language Dataset - YounghoonJeongKAIST (Korea Advanced Institute of Science and Technology) + YounghunJeongKAIST (Korea Advanced Institute of Science and Technology) JuhyunOhIndependent Researcher JongwonLeeSamsung Research JaimeenAhnIndependent Researcher From 6e28a580fe5778df292482888f83fd47ee210d25 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 19:37:39 -0600 Subject: [PATCH 10/69] Paper meta data correction for 2023.emnlp-demo.13, closes #2900. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index fab8731591..00f813476b 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -12860,7 +12860,7 @@ SukannyaPurkayasthaTU Darmstadt LeonEngländerTechnical University of Darmstadt TimoImhofTechnical University of Darmstadt - IvanVuliUniversity of Cambridge + IvanVulićUniversity of Cambridge SebastianRuderGoogle IrynaGurevychUKP Lab, Technische Universität Darmstadt JonasPfeifferGoogle From b0672d997f066f3b7245cb77931b8f6176c98ede Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sat, 9 Dec 2023 20:14:58 -0600 Subject: [PATCH 11/69] Paper pdf correction for 2023.nllp-1.10, closes #2901. --- data/xml/2023.nllp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.nllp.xml b/data/xml/2023.nllp.xml index 46e0caa19c..487f1813f2 100644 --- a/data/xml/2023.nllp.xml +++ b/data/xml/2023.nllp.xml @@ -127,8 +127,10 @@ LakshminarayananSubramanianNew York University 85-98 This paper formulates a new task of extracting privacy parameters from a privacy policy, through the lens of Contextual Integrity (CI), an established social theory framework for reasoning about privacy norms. Through extensive experiments, we further show that incorporating CI-based domain-specific knowledge into a BERT-based SRL model results in the highest precision and recall, achieving an F1 score of 84%. With our work, we would like to motivate new research in building NLP applications for the privacy domain. - 2023.nllp-1.10 + 2023.nllp-1.10 shvartzshanider-etal-2023-beyond + + Typo correction in the CI labels in Section 5.4. Towards Mitigating Perceived Unfairness in Contracts from a Non-Legal Stakeholder’s Perspective From 41a693bbd229770a0a9a0cf2cba09a0777eb1c18 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sun, 10 Dec 2023 20:22:57 -0600 Subject: [PATCH 12/69] Paper pdf correction for 2020.emnlp-main.254, closes #2902. --- data/xml/2020.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2020.emnlp.xml b/data/xml/2020.emnlp.xml index f1b10e5845..b637ab1aaa 100644 --- a/data/xml/2020.emnlp.xml +++ b/data/xml/2020.emnlp.xml @@ -3819,11 +3819,13 @@ RyanCotterell 3138–3153 The question of how to probe contextual word representations in a way that is principled and useful has seen significant recent attention. In our contribution to this discussion, we argue, first, for a probe metric that reflects the trade-off between probe complexity and performance: the Pareto hypervolume. To measure complexity, we present a number of parametric and non-parametric metrics. Our experiments with such metrics show that probe’s performance curves often fail to align with widely accepted rankings between language representations (with, e.g., non-contextual representations outperforming contextual ones). These results lead us to argue, second, that common simplistic probe tasks such as POS labeling and dependency arc labeling, are inadequate to evaluate the properties encoded in contextual word representations. We propose full dependency parsing as an example probe task, and demonstrate it with the Pareto hypervolume. In support of our arguments, the results of this illustrative experiment conform closer to accepted rankings among contextual word representations. - 2020.emnlp-main.254 + 2020.emnlp-main.254 10.18653/v1/2020.emnlp-main.254 Interpretation of <fixed-case>NLP</fixed-case> models through input marginalization From 33ee9c1c45c404699310fb4910666589bfdcaffb Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sun, 10 Dec 2023 20:33:26 -0600 Subject: [PATCH 13/69] Paper meta data correction for 2022.emnlp-main.788, closes #2903. --- data/xml/2022.emnlp.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2022.emnlp.xml b/data/xml/2022.emnlp.xml index 63ef3add94..a301634db9 100644 --- a/data/xml/2022.emnlp.xml +++ b/data/xml/2022.emnlp.xml @@ -10260,13 +10260,13 @@ 10.18653/v1/2022.emnlp-main.787 - Attentional Probe: Estimating a Module’s Functional Potential + The Architectural Bottleneck Principle TiagoPimentelUniversity of Cambridge JosefValvodaUniversity of Cambridge NiklasStoehrETH Zurich RyanCotterellETH Zürich 11459-11472 - + In this paper, we seek to measure how much information a component in a neural network could extract from the representations fed into it. Our work stands in contrast to prior probing work, most of which investigates how much information a model's representations contain. This shift in perspective leads us to propose a new principle for probing, the architectural bottleneck principle: In order to estimate how much information a given component could extract, a probe should look exactly like the component. Relying on this principle, we estimate how much syntactic information is available to transformers through our attentional probe, a probe that exactly resembles a transformer's self-attention head. Experimentally, we find that, in three models (BERT, ALBERT, and RoBERTa), a sentence's syntax tree is mostly extractable by our probe, suggesting these models have access to syntactic information while composing their contextual representations. Whether this information is actually used by these models, however, remains an open question.. 2022.emnlp-main.788 pimentel-etal-2022-attentional 10.18653/v1/2022.emnlp-main.788 From e6c1bfdb6f00ae9217718ce4375ce03847302869 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sun, 10 Dec 2023 20:38:23 -0600 Subject: [PATCH 14/69] Paper meta data correction for 2023.eamt-1.55, closes #2909. --- data/xml/2023.eamt.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.eamt.xml b/data/xml/2023.eamt.xml index f872fef12c..e9f324d5d1 100644 --- a/data/xml/2023.eamt.xml +++ b/data/xml/2023.eamt.xml @@ -637,7 +637,7 @@ <fixed-case>M</fixed-case>a<fixed-case>C</fixed-case>o<fixed-case>C</fixed-case>u: Massive collection and curation of monolingual and bilingual data: focus on under-resourced languages - Marta BaNón + MartaBañón MălinaChichirău MiquelEsplà-Gomis MikelForcada From 54edea1cce83d2eaecc6bbcad4878acd8c789fb0 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sun, 10 Dec 2023 20:41:25 -0600 Subject: [PATCH 15/69] Paper mata data correction for 2023.banglalp-1.5, closes #2910. --- data/xml/2023.banglalp.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.banglalp.xml b/data/xml/2023.banglalp.xml index 1d49442547..3a0deb7c4f 100644 --- a/data/xml/2023.banglalp.xml +++ b/data/xml/2023.banglalp.xml @@ -72,8 +72,8 @@ SourabrataMukherjee AkankshaBansal PrithaMajumdar - AtulOjhaUniversity of Galway, Ireland, Insight SFI Research Centre for Data Analytics, DSI, University of Galway, Ireland and Panlingua Languague Processing LLP, India - OndrejDusekCharles University, Prague + Atul Kr.OjhaUniversity of Galway, Ireland, Insight SFI Research Centre for Data Analytics, DSI, University of Galway, Ireland and Panlingua Languague Processing LLP, India + OndřejDušekCharles University, Prague 34-47 Text style transfer (TST) involves modifying the linguistic style of a given text while retaining its core content. This paper addresses the challenging task of text style transfer in the Bangla language, which is low-resourced in this area. We present a novel Bangla dataset that facilitates text sentiment transfer, a subtask of TST, enabling the transformation of positive sentiment sentences to negative and vice versa. To establish a high-quality base for further research, we refined and corrected an existing English dataset of 1,000 sentences for sentiment transfer based on Yelp reviews, and we introduce a new human-translated Bangla dataset that parallels its English counterpart. Furthermore, we offer multiple benchmark models that serve as a validation of the dataset and baseline for further research. 2023.banglalp-1.5 From 4965e0cb48c8a66e973dbce2de3a65aa8a349e06 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sun, 10 Dec 2023 20:43:43 -0600 Subject: [PATCH 16/69] Paper meta data correction for 2023.banglalp-1.27, closes #2911. --- data/xml/2023.banglalp.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.banglalp.xml b/data/xml/2023.banglalp.xml index 3a0deb7c4f..3b74c191c7 100644 --- a/data/xml/2023.banglalp.xml +++ b/data/xml/2023.banglalp.xml @@ -330,8 +330,8 @@ <fixed-case>UFAL</fixed-case>-<fixed-case>ULD</fixed-case> at <fixed-case>BLP</fixed-case>-2023 Task 1: Violence Detection in <fixed-case>B</fixed-case>angla Text SourabrataMukherjee - AtulOjhaUniversity of Galway, Ireland, Insight SFI Research Centre for Data Analytics, DSI, University of Galway, Ireland and Panlingua Languague Processing LLP, India - OndrejDusekCharles University, Prague + Atul Kr.OjhaUniversity of Galway, Ireland, Insight SFI Research Centre for Data Analytics, DSI, University of Galway, Ireland and Panlingua Languague Processing LLP, India + OndřejDušekCharles University, Prague 220-224 In this paper, we present UFAL-ULD team’s system, desinged as a part of the BLP Shared Task 1: Violence Inciting Text Detection (VITD). This task aims to classify text, with a particular challenge of identifying incitement to violence into Direct, Indirect or Non-violence levels. We experimented with several pre-trained sequence classification models, including XLM-RoBERTa, BanglaBERT, Bangla BERT Base, and Multilingual BERT. Our best-performing model was based on the XLM-RoBERTa-base architecture, which outperformed the baseline models. Our system was ranked 20th among the 27 teams that participated in the task. 2023.banglalp-1.27 From 0e6b304854297cb85a048f6abfa5315d3c5a71de Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Sun, 10 Dec 2023 20:55:18 -0600 Subject: [PATCH 17/69] Paper meta data correction for 2023.banglalp-1.45, closes #2912. --- data/xml/2023.banglalp.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.banglalp.xml b/data/xml/2023.banglalp.xml index 3b74c191c7..36e71f9eab 100644 --- a/data/xml/2023.banglalp.xml +++ b/data/xml/2023.banglalp.xml @@ -527,8 +527,8 @@ <fixed-case>UFAL</fixed-case>-<fixed-case>ULD</fixed-case> at <fixed-case>BLP</fixed-case>-2023 Task 2 Sentiment Classification in <fixed-case>B</fixed-case>angla Text SourabrataMukherjee - AtulOjhaUniversity of Galway, Ireland, Insight SFI Research Centre for Data Analytics, DSI, University of Galway, Ireland and Panlingua Languague Processing LLP, India - OndrejDusekCharles University, Prague + Atul Kr.OjhaUniversity of Galway, Ireland, Insight SFI Research Centre for Data Analytics, DSI, University of Galway, Ireland and Panlingua Languague Processing LLP, India + OndřejDušekCharles University, Prague 336-339 In this paper, we present the UFAL-ULD team’s system for the BLP Shared Task 2: Sentiment Analysis of Bangla Social Media Posts. The Task 2 involves classifying text into Positive, Negative, or Neutral sentiments. As a part of this task, we conducted a series of experiments with several pre-trained sequence classification models – XLM-RoBERTa, BanglaBERT, Bangla BERT Base and Multilingual BERT. Among these, our best-performing model was based on the XLM-RoBERTa-base architecture, which outperforms baseline models. Our system was ranked 19th among the 30 teams that participated in the task. 2023.banglalp-1.45 From b3072852a414cd0e60544ff1bce3058740726ef0 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:19:37 -0600 Subject: [PATCH 18/69] Paper meta data correction for 2023.emnlp-main.920, closes #2915. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 00f813476b..b586b64a20 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -11036,7 +11036,7 @@ RidwanMahbub IfradKhan SamihaAnuva - MdShahriar + Md ShihabShahriar Md Tahmid RahmanLaskar SabbirAhmed 14878-14886 From 9620b12715d2af027fd8007506194820a99e2b70 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:22:42 -0600 Subject: [PATCH 19/69] Paper meta data correction for 2023.findings-emnlp.413, closes #2917. --- data/xml/2023.findings.xml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index f6cb9f22c8..5f7a4214ca 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -19284,10 +19284,10 @@ Responsible <fixed-case>AI</fixed-case> Considerations in Text Summarization Research: A Review of Current Practices - YuLiu + Yu LuLiu MengCao - SuBlodgett - JackieCheung + Su LinBlodgett + Jackie Chi KitCheung AlexandraOlteanu AdamTrischler 6246-6261 From 000dd25caa3b56ae9243b44314b0d738721bb0dd Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:24:43 -0600 Subject: [PATCH 20/69] Paper meta data correction for 2023.arabicnlp-1.28, closes #2918. --- data/xml/2023.arabicnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.arabicnlp.xml b/data/xml/2023.arabicnlp.xml index 631aa03d2e..64d9435483 100644 --- a/data/xml/2023.arabicnlp.xml +++ b/data/xml/2023.arabicnlp.xml @@ -336,7 +336,7 @@ In-Context Meta-Learning vs. Semantic Score-Based Similarity: A Comparative Study in <fixed-case>A</fixed-case>rabic Short Answer Grading MennaFateen - TsunenoriMina + TsunenoriMine 350-358 Delegating short answer grading to automated systems enhances efficiency, giving teachers more time for vital human-centered aspects of education. Studies in automatic short answer grading (ASAG) approach the problem from instance-based or reference-based perspectives. Recent studies have favored instance-based methods, but they demand substantial data for training, which is often scarce in classroom settings. This study compares both approaches using an Arabic ASAG dataset. We employ in-context meta-learning for instance-based and semantic score-based similarity for reference-based grading. Results show both methods outperform a baseline and occasionally even surpass human raters when grading unseen answers. Notably, the semantic score-based similarity approach excels in zero-shot settings, outperforming in-context meta-learning. Our work contributes insights to Arabic ASAG and introduces a prompt category classification model, leveraging GPT3.5 to augment Arabic data for improved performance. 2023.arabicnlp-1.28 From db06b6e9c7edafa2ae69f3269c894ea0bba31f63 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:29:12 -0600 Subject: [PATCH 21/69] Paper meta data correction for 2023.findings-emnlp.841, closes #2924. --- data/xml/2023.findings.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 5f7a4214ca..a161b34f28 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -24361,8 +24361,8 @@ <fixed-case>MISCA</fixed-case>: A Joint Model for Multiple Intent Detection and Slot Filling with Intent-Slot Co-Attention ThinhPham - TranChi - DatNguyen + ChiTran + Dat QuocNguyen 12641-12650 The research study of detecting multiple intents and filling slots is becoming more popular because of its relevance to complicated real-world situations. Recent advanced approaches, which are joint models based on graphs, might still face two potential issues: (i) the uncertainty introduced by constructing graphs based on preliminary intents and slots, which may transfer intent-slot correlation information to incorrect label node destinations, and (ii) direct incorporation of multiple intent labels for each token w.r.t. token-level intent voting might potentially lead to incorrect slot predictions, thereby hurting the overall performance. To address these two issues, we propose a joint model named MISCA. Our MISCA introduces an intent-slot co-attention mechanism and an underlying layer of label attention mechanism. These mechanisms enable MISCA to effectively capture correlations between intents and slot labels, eliminating the need for graph construction. They also facilitate the transfer of correlation information in both directions: from intents to slots and from slots to intents, through multiple levels of label-specific representations, without relying on token-level intent information. Experimental results show that MISCA outperforms previous models, achieving new state-of-the-art overall accuracy performances on two benchmark datasets MixATIS and MixSNIPS. This highlights the effectiveness of our attention mechanisms. 2023.findings-emnlp.841 From 2a7aa14494d53bdc42322737674d53fa1d804298 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:31:07 -0600 Subject: [PATCH 22/69] Paper meta data correction for 2023.emnlp-main.388, closes #2925. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index b586b64a20..34fc1981dc 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -4688,7 +4688,7 @@ <fixed-case>B</fixed-case>asaha<fixed-case>C</fixed-case>orpus: An Expanded Linguistic Resource for Readability Assessment in <fixed-case>C</fixed-case>entral <fixed-case>P</fixed-case>hilippine Languages - JosephImperial + Joseph MarvinImperial EkaterinaKochmar 6302-6309 Current research on automatic readability assessment (ARA) has focused on improving the performance of models in high-resource languages such as English. In this work, we introduce and release BasahaCorpus as part of an initiative aimed at expanding available corpora and baseline models for readability assessment in lower resource languages in the Philippines. We compiled a corpus of short fictional narratives written in Hiligaynon, Minasbate, Karay-a, and Rinconada—languages belonging to the Central Philippine family tree subgroup—to train ARA models using surface-level, syllable-pattern, and n-gram overlap features. We also propose a new hierarchical cross-lingual modeling approach that takes advantage of a language’s placement in the family tree to increase the amount of available training data. Our study yields encouraging results that support previous work showcasing the efficacy of cross-lingual models in low-resource settings, as well as similarities in highly informative linguistic features for mutually intelligible languages. From 99231d77d9266ac1f4c11b4113126ee184bdc458 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:34:03 -0600 Subject: [PATCH 23/69] Paper meta data correction for 2023.findings-emnlp.805, closes #2926. --- data/xml/2023.findings.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index a161b34f28..a2784faa3f 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -23904,8 +23904,8 @@ Uniform Complexity for Text Generation - JosephImperial - HarishMadabushi + Joseph MarvinImperial + Harish TayyarMadabushi 12025-12046 Large language models (LLMs) have shown promising results in a wide array of generative NLP tasks, such as summarization and machine translation. In the context of narrative generation, however, existing models still do not capture factors that contribute to producing consistent text. For instance, it is logical that a piece of text or a story should be uniformly readable throughout and that this form of complexity should be controllable. As such, if the complexity of an input text prompt is rated first-grade reading level in the Flesch Reading Ease test, then the generated text continuing the plot should also be within this range of complexity. With this in mind, we introduce Uniform Complexity for Text Generation (UCTG), a new benchmark test which raises the challenge of making generative models observe uniform linguistic properties with respect to prompts. We experiment with over 150+ linguistically and cognitively motivated features for evaluating text complexity in humans and generative models. From our results, we find that models such as GPT-2 struggle to preserve the complexity of input prompts used in its generations, even if finetuned with professionally written texts. 2023.findings-emnlp.805 From 61a80fc604c15a339e3fb874764b023bb6fc7aa8 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:35:29 -0600 Subject: [PATCH 24/69] Paper pdf correction for 2023.emnlp-main.167, closes #2932. --- data/xml/2023.emnlp.xml | 13 ++++++------- 1 file changed, 6 insertions(+), 7 deletions(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 34fc1981dc..066f50d305 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -2037,8 +2037,10 @@ HarksooKim 2789-2799 Most research on multimodal open-domain dialogue agents has focused on pretraining and multi-task learning using additional rich datasets beyond a given target dataset. However, methods for exploiting these additional datasets can be quite limited in real-world settings, creating a need for more efficient methods for constructing agents based solely on the target dataset. To address these issues, we present a new learning strategy called vision-language warm-up tasks for multimodal dialogue models (VLAW-MDM). This strategy does not require the use of large pretraining or multi-task datasets but rather relies solely on learning from target data. Moreover, our proposed approach automatically generate captions for images and incorporate them into the model’s input to improve the contextualization of visual information. Using this novel approach, we empirically demonstrate that our learning strategy is effective for limited data and relatively small models. The result show that our method achieved comparable and in some cases superior performance compared to existing state-of-the-art models on various evaluation metrics. - 2023.emnlp-main.167 + 2023.emnlp-main.167 lee-etal-2023-framework + + Fixed the sponsor in the Acknowledgments section. Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair Modeling @@ -6871,8 +6873,7 @@ DavidStap ChristofMonz 9200-9208 - - k-nearest-neighbor machine translation has demonstrated remarkable improvements in machine translation quality by creating a datastore of cached examples. However, these improvements have been limited to high-resource language pairs, with large datastores, and remain a challenge for low-resource languages. In this paper, we address this issue by combining representations from multiple languages into a single datastore. Our results consistently demonstrate substantial improvements not only in low-resource translation quality (up to +3.6 BLEU), but also for high-resource translation quality (up to +0.5 BLEU). Our experiments show that it is possible to create multilingual datastores that are a quarter of the size, achieving a 5.3x speed improvement, by using linguistic similarities for datastore creation. + k-nearest-neighbor machine translation has demonstrated remarkable improvements in machine translation quality by creating a datastore of cached examples. However, these improvements have been limited to high-resource language pairs, with large datastores, and remain a challenge for low-resource languages. In this paper, we address this issue by combining representations from multiple languages into a single datastore. Our results consistently demonstrate substantial improvements not only in low-resource translation quality (up to +3.6 BLEU), but also for high-resource translation quality (up to +0.5 BLEU). Our experiments show that it is possible to create multilingual datastores that are a quarter of the size, achieving a 5.3x speed improvement, by using linguistic similarities for datastore creation. 2023.emnlp-main.571 stap-monz-2023-multilingual @@ -12784,8 +12785,7 @@ LeZhangFuxi AI Lab, NetEase Inc. JiashuPuNetEase Fuxi Lab 99-105 - Lyrics generation is a well-known application in natural language generation research, with several previous studies focusing on generating accurate lyrics using precise control such as keywords, rhymes, etc. However, lyrics imitation, which involves writing new lyrics by imitating the style and content of the source lyrics, remains a challenging task due to the lack of a parallel corpus. In this paper, we introduce Sudowoodo, a Chinese lyrics imitation system that can generate new lyrics based on the text of source lyrics. To address the issue of lacking a parallel training corpus for lyrics imitation, we propose a novel framework to construct a parallel corpus based on a keyword-based lyrics model from source lyrics. Then the pairs (new lyrics, source lyrics) are used to train the lyrics imitation model. During the inference process, we utilize a post-processing module to filter and rank the generated lyrics, selecting the highest-quality ones. We incorporated audio information and aligned the lyrics with the audio to form the songs as a bonus. The human evaluation results show that our framework can perform better lyric imitation. Meanwhile, the Sudowoodo system and demo video of the system is available at Sudowoodo and https://youtu.be/u5BBT\_j1L5M - + Lyrics generation is a well-known application in natural language generation research, with several previous studies focusing on generating accurate lyrics using precise control such as keywords, rhymes, etc. However, lyrics imitation, which involves writing new lyrics by imitating the style and content of the source lyrics, remains a challenging task due to the lack of a parallel corpus. In this paper, we introduce Sudowoodo, a Chinese lyrics imitation system that can generate new lyrics based on the text of source lyrics. To address the issue of lacking a parallel training corpus for lyrics imitation, we propose a novel framework to construct a parallel corpus based on a keyword-based lyrics model from source lyrics. Then the pairs (new lyrics, source lyrics) are used to train the lyrics imitation model. During the inference process, we utilize a post-processing module to filter and rank the generated lyrics, selecting the highest-quality ones. We incorporated audio information and aligned the lyrics with the audio to form the songs as a bonus. The human evaluation results show that our framework can perform better lyric imitation. Meanwhile, the Sudowoodo system and demo video of the system is available at Sudowoodo and https://youtu.be/u5BBT\_j1L5M 2023.emnlp-demo.8 chang-etal-2023-sudowoodo @@ -13280,8 +13280,7 @@ KangLiuInstitute of Automation, Chinese Academy of Sciences JunZhaoNLPR, Institute of Automation, Chinese Academy of Sciences 479-494 - The unprecedented performance of LLMs requires comprehensive and accurate evaluation. We argue that for LLMs evaluation, benchmarks need to be comprehensive and systematic. To this end, we propose the Zhujiu benchmark, which has the following strengths: (1) Multi-dimensional ability coverage: We comprehensively evaluate LLMs across 7 ability dimensions covering 51 tasks. Especially, we also propose a new benchmark that focus on knowledge ability of LLMs. (2) Multi-faceted evaluation methods collaboration: We use 3 different yet complementary evaluation methods to comprehensively evaluate LLMs, which can ensure the authority and accuracy of the evaluation results. (3) Comprehensive Chinese benchmark: ZhuJiu is the pioneering benchmark that fully assesses LLMs in Chinese, while also providing equally robust evaluation abilities in English. (4) Avoiding potential data leakage: To avoid data leakage, we construct evaluation data specifically for 37 tasks. We evaluate 10 current mainstream LLMs, and conduct an in-depth discussion and analysis of their results. The ZhuJiu benchmark and open-participation leaderboard are publicly released at http://www.zhujiu-benchmark.com and we also provide a demo video at https://youtu.be/qypkJ89L1Ic. - + The unprecedented performance of LLMs requires comprehensive and accurate evaluation. We argue that for LLMs evaluation, benchmarks need to be comprehensive and systematic. To this end, we propose the Zhujiu benchmark, which has the following strengths: (1) Multi-dimensional ability coverage: We comprehensively evaluate LLMs across 7 ability dimensions covering 51 tasks. Especially, we also propose a new benchmark that focus on knowledge ability of LLMs. (2) Multi-faceted evaluation methods collaboration: We use 3 different yet complementary evaluation methods to comprehensively evaluate LLMs, which can ensure the authority and accuracy of the evaluation results. (3) Comprehensive Chinese benchmark: ZhuJiu is the pioneering benchmark that fully assesses LLMs in Chinese, while also providing equally robust evaluation abilities in English. (4) Avoiding potential data leakage: To avoid data leakage, we construct evaluation data specifically for 37 tasks. We evaluate 10 current mainstream LLMs, and conduct an in-depth discussion and analysis of their results. The ZhuJiu benchmark and open-participation leaderboard are publicly released at http://www.zhujiu-benchmark.com and we also provide a demo video at https://youtu.be/qypkJ89L1Ic. 2023.emnlp-demo.44 zhang-etal-2023-zhujiu From 9755d51d48524247e3a6bffc57bf1bc6aec719db Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Mon, 11 Dec 2023 20:37:25 -0600 Subject: [PATCH 25/69] Paper pdf correction for 2023.findings-emnlp.459, closes #2933. --- data/xml/2023.findings.xml | 51 ++++++++++++++------------------------ 1 file changed, 19 insertions(+), 32 deletions(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index a2784faa3f..53800674be 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -137,8 +137,7 @@ 2023.findings-eacl.9 2023.findings-eacl.9.dataset.zip roychowdhury-gupta-2023-data - Scarcity of large-scale datasets, especially for resource-impoverished languages motivates exploration of data-efficient methods for hate speech detection. Hateful intents are expressed explicitly (use of cuss, swear, abusive words) and implicitly (indirect and contextual). In this work, we progress implicit and explicit hate speech detection using an input-level data augmentation technique, task reformulation using entailment and cross-learning across five languages. Our proposed data augmentation technique EasyMix, improves the performance across all english datasets by ~1% and across multilingual datasets by ~1-9%. We also observe substantial gains of ~2-8% by reformulating hate speech detection as entail problem. We further probe the contextual models and observe that higher layers encode implicit hate while lower layers focus on explicit hate, highlighting the importance of token-level understanding for explicit and context-level for implicit hate speech detection. Code and Dataset splits - https://anonymous.4open.science/r/data_efficient_hatedetect/ - + Scarcity of large-scale datasets, especially for resource-impoverished languages motivates exploration of data-efficient methods for hate speech detection. Hateful intents are expressed explicitly (use of cuss, swear, abusive words) and implicitly (indirect and contextual). In this work, we progress implicit and explicit hate speech detection using an input-level data augmentation technique, task reformulation using entailment and cross-learning across five languages. Our proposed data augmentation technique EasyMix, improves the performance across all english datasets by ~1% and across multilingual datasets by ~1-9%. We also observe substantial gains of ~2-8% by reformulating hate speech detection as entail problem. We further probe the contextual models and observe that higher layers encode implicit hate while lower layers focus on explicit hate, highlighting the importance of token-level understanding for explicit and context-level for implicit hate speech detection. Code and Dataset splits - https://anonymous.4open.science/r/data_efficient_hatedetect/ @@ -1488,8 +1487,7 @@ RichardJohanssonUniversity of Gothenburg MarcoKuhlmannLinköping University 1485-1493 - Recent work on the Retrieval-Enhanced Transformer (RETRO) model has shown impressive results: off-loading memory from trainable weights to a retrieval database can significantly improve language modeling and match the performance of non-retrieval models that are an order of magnitude larger in size. It has been suggested that at least some of this performance gain is due to non-trivial generalization based on both model weights and retrieval. In this paper, we try to better understand the relative contributions of these two components. We find that the performance gains from retrieval to a very large extent originate from overlapping tokens between the database and the test data, suggesting less of non-trivial generalization than previously assumed. More generally, our results point to the challenges of evaluating the generalization of retrieval-augmented language models such as RETRO, as even limited token overlap may significantly decrease test-time loss. We release our code and model at https://github.com/TobiasNorlund/retro - + Recent work on the Retrieval-Enhanced Transformer (RETRO) model has shown impressive results: off-loading memory from trainable weights to a retrieval database can significantly improve language modeling and match the performance of non-retrieval models that are an order of magnitude larger in size. It has been suggested that at least some of this performance gain is due to non-trivial generalization based on both model weights and retrieval. In this paper, we try to better understand the relative contributions of these two components. We find that the performance gains from retrieval to a very large extent originate from overlapping tokens between the database and the test data, suggesting less of non-trivial generalization than previously assumed. More generally, our results point to the challenges of evaluating the generalization of retrieval-augmented language models such as RETRO, as even limited token overlap may significantly decrease test-time loss. We release our code and model at https://github.com/TobiasNorlund/retro 2023.findings-eacl.109 norlund-etal-2023-generalization @@ -19093,8 +19081,7 @@ DangyangChen JixiongChen 5933-5957 - Personalized dialogue systems aim to endow the chatbot agent with more anthropomorphic traits for human-like interactions. Previous approaches have explored explicitly user profile modeling using text descriptions, implicit derivation of user embeddings, or utilizing handicraft prompts for ChatGPT-like models. However, textual personas are limited in describing multi-faceted attributes (e.g., language style, inner character nuances), implicit embedding suffers from personality sparsity, and handicraft prompts lack fine-grained and stable controllability. Hence, these approaches may struggle with complex personalized dialogue generation tasks that require generating controllable responses with multiple personal attributes. To this end, we propose Miracle, a novel personalized dialogue generation method through MultIple PeRsonal Attributes Control within Latent-Space Energy-based Models. ttributes Control within Latent-Space Energy-based Models. Specifically, our approach first disentangles complex personality into multi-faceted attributes. Subsequently, we employ a conditional variational auto-encoder to align with the dense personalized responses within a latent joint attribute space. We have also tailored a dedicated energy function and customized the ordinary differential equations sampling method to offer flexible attribute composition and precise attribute control. Extensive experiments demonstrate that Miracle outperforms several strong baselines in terms of personality controllability and response generation quality. Our dataset and code are available at https://github.com/LZY-the-boys/MIRACLE - + Personalized dialogue systems aim to endow the chatbot agent with more anthropomorphic traits for human-like interactions. Previous approaches have explored explicitly user profile modeling using text descriptions, implicit derivation of user embeddings, or utilizing handicraft prompts for ChatGPT-like models. However, textual personas are limited in describing multi-faceted attributes (e.g., language style, inner character nuances), implicit embedding suffers from personality sparsity, and handicraft prompts lack fine-grained and stable controllability. Hence, these approaches may struggle with complex personalized dialogue generation tasks that require generating controllable responses with multiple personal attributes. To this end, we propose Miracle, a novel personalized dialogue generation method through MultIple PeRsonal Attributes Control within Latent-Space Energy-based Models. ttributes Control within Latent-Space Energy-based Models. Specifically, our approach first disentangles complex personality into multi-faceted attributes. Subsequently, we employ a conditional variational auto-encoder to align with the dense personalized responses within a latent joint attribute space. We have also tailored a dedicated energy function and customized the ordinary differential equations sampling method to offer flexible attribute composition and precise attribute control. Extensive experiments demonstrate that Miracle outperforms several strong baselines in terms of personality controllability and response generation quality. Our dataset and code are available at https://github.com/LZY-the-boys/MIRACLE 2023.findings-emnlp.395 lu-etal-2023-miracle @@ -19833,8 +19820,10 @@ WenjunKe 6877-6892 Relation extraction (RE) consistently involves a certain degree of labeled or unlabeled data even if under zero-shot setting. Recent studies have shown that large language models (LLMs) transfer well to new tasks out-of-the-box simply given a natural language prompt, which provides the possibility of extracting relations from text without any data and parameter tuning. This work focuses on the study of exploring LLMs, such as ChatGPT, as zero-shot relation extractors. On the one hand, we analyze the drawbacks of existing RE prompts and attempt to incorporate recent prompt techniques such as chain-of-thought (CoT) to improve zero-shot RE. We propose the summarize-and-ask (SumAsk) prompting, a simple prompt recursively using LLMs to transform RE inputs to the effective question answering (QA) format. On the other hand, we conduct comprehensive experiments on various benchmarks and settings to investigate the capabilities of LLMs on zero-shot RE. Specifically, we have the following findings: (i) SumAsk consistently and significantly improves LLMs performance on different model sizes, benchmarks and settings; (ii) Zero-shot prompting with ChatGPT achieves competitive or superior results compared with zero-shot and fully supervised methods; (iii) LLMs deliver promising performance in extracting overlapping relations; (iv) The performance varies greatly regarding different relations. Different from small language models, LLMs are effective in handling challenge none-of-the-above (NoTA) relation. - 2023.findings-emnlp.459 + 2023.findings-emnlp.459 li-etal-2023-revisiting-large + + Updated Acknowledgement. Multi-Stage Pre-training Enhanced by <fixed-case>C</fixed-case>hat<fixed-case>GPT</fixed-case> for Multi-Scenario Multi-Domain Dialogue Summarization @@ -23119,8 +23108,7 @@ AkshatShrivastava AleksandrLivshits 11082-11092 - - Autoregressive (AR) encoder-decoder neural networks have proved successful in many NLP problems, including Semantic Parsing – a task that translates natural language to machine-readable parse trees. However, the sequential prediction process of AR models can be slow. To accelerate AR for semantic parsing, we introduce a new technique called TreePiece that tokenizes a parse tree into subtrees and generates one subtree per decoding step. On TOPv2 benchmark, TreePiece shows 4.6 times faster decoding speed than standard AR, and comparable speed but significantly higher accuracy compared to Non-Autoregressive (NAR). + Autoregressive (AR) encoder-decoder neural networks have proved successful in many NLP problems, including Semantic Parsing – a task that translates natural language to machine-readable parse trees. However, the sequential prediction process of AR models can be slow. To accelerate AR for semantic parsing, we introduce a new technique called TreePiece that tokenizes a parse tree into subtrees and generates one subtree per decoding step. On TOPv2 benchmark, TreePiece shows 4.6 times faster decoding speed than standard AR, and comparable speed but significantly higher accuracy compared to Non-Autoregressive (NAR). 2023.findings-emnlp.740 wang-etal-2023-treepiece @@ -25232,8 +25220,7 @@ FoscaGiannotti DongwonLee 13706-13724 - - \textit{Authorship Analysis}, also known as stylometry, has been an essential aspect of Natural Language Processing (NLP) for a long time. Likewise, the recent advancement of Large Language Models (LLMs) has made authorship analysis increasingly crucial for distinguishing between human-written and AI-generated texts. However, these authorship analysis tasks have primarily been focused on \textit{written texts}, not considering \textit{spoken texts}. Thus, we introduce the largest benchmark for spoken texts - {\sf HANSEN}(\underline{H}uman \underline{AN}d ai \underline{S}poken t\underline{E}xt be\underline{N}chmark). {\sf HANSEN} encompasses meticulous curation of existing speech datasets accompanied by transcripts, alongside the creation of novel AI-generated spoken text datasets. Together, it comprises 17 human datasets, and AI-generated spoken texts created using 3 prominent LLMs: ChatGPT, PaLM2, and Vicuna13B. To evaluate and demonstrate the utility of {\sf HANSEN}, we perform Authorship Attribution (AA) & Author Verification (AV) on human-spoken datasets and conducted Human vs. AI text detection using state-of-the-art (SOTA) models. While SOTA methods, such as, character n-gram or Transformer-based model, exhibit similar AA & AV performance in human-spoken datasets compared to written ones, there is much room for improvement in AI-generated spoken text detection. The {\sf HANSEN} benchmark is available at: https://huggingface.co/datasets/HANSEN-REPO/HANSEN + \textit{Authorship Analysis}, also known as stylometry, has been an essential aspect of Natural Language Processing (NLP) for a long time. Likewise, the recent advancement of Large Language Models (LLMs) has made authorship analysis increasingly crucial for distinguishing between human-written and AI-generated texts. However, these authorship analysis tasks have primarily been focused on \textit{written texts}, not considering \textit{spoken texts}. Thus, we introduce the largest benchmark for spoken texts - {\sf HANSEN}(\underline{H}uman \underline{AN}d ai \underline{S}poken t\underline{E}xt be\underline{N}chmark). {\sf HANSEN} encompasses meticulous curation of existing speech datasets accompanied by transcripts, alongside the creation of novel AI-generated spoken text datasets. Together, it comprises 17 human datasets, and AI-generated spoken texts created using 3 prominent LLMs: ChatGPT, PaLM2, and Vicuna13B. To evaluate and demonstrate the utility of {\sf HANSEN}, we perform Authorship Attribution (AA) & Author Verification (AV) on human-spoken datasets and conducted Human vs. AI text detection using state-of-the-art (SOTA) models. While SOTA methods, such as, character n-gram or Transformer-based model, exhibit similar AA & AV performance in human-spoken datasets compared to written ones, there is much room for improvement in AI-generated spoken text detection. The {\sf HANSEN} benchmark is available at: https://huggingface.co/datasets/HANSEN-REPO/HANSEN 2023.findings-emnlp.916 tripto-etal-2023-hansen From a4256688112504073e4bafdabb70cf682fec4f58 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 20:49:49 -0600 Subject: [PATCH 26/69] Meta data correction for 2023.emnlp-main.128, closes #2928. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 066f50d305..78118073c9 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -1548,7 +1548,7 @@ <fixed-case>V</fixed-case>ives<fixed-case>D</fixed-case>ebate-Speech: A Corpus of Spoken Argumentation to Leverage Audio Features for Argument Mining RamonRuiz-Dolz - JavierSanchez + JavierIranzo-Sánchez 2071-2077 In this paper, we describe VivesDebate-Speech, a corpus of spoken argumentation created to leverage audio features for argument mining tasks. The creation of this corpus represents an important contribution to the intersection of speech processing and argument mining communities, and one of the most complete publicly available resources in this topic. Moreover, we have performed a set of first-of-their-kind experiments which show an improvement when integrating audio features into the argument mining pipeline. The provided results can be used as a baseline for future research. 2023.emnlp-main.128 From d26d8c769ac0b8395de8c3181b3db4afa9c5c706 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 20:52:38 -0600 Subject: [PATCH 27/69] Paper meta data correction forr 2022.acl-long.480, closes #2929. --- data/xml/2022.acl.xml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/data/xml/2022.acl.xml b/data/xml/2022.acl.xml index 7c1bb8fffa..10df23146b 100644 --- a/data/xml/2022.acl.xml +++ b/data/xml/2022.acl.xml @@ -7696,11 +7696,11 @@ in the Case of Unambiguous Gender From Simultaneous to Streaming Machine Translation by Leveraging Streaming History - JavierIranzo Sanchez + JavierIranzo-Sánchez JorgeCivera - AlfonsJuan-Císcar + AlfonsJuan 6972-6985 - Simultaneous machine translation has recently gained traction thanks to significant quality improvements and the advent of streaming applications. Simultaneous translation systems need to find a trade-off between translation quality and response time, and with this purpose multiple latency measures have been proposed. However, latency evaluations for simultaneous translation are estimated at the sentence level, not taking into account the sequential nature of a streaming scenario. Indeed, these sentence-level latency measures are not well suited for continuous stream translation, resulting in figures that are not coherent with the simultaneous translation policy of the system being assessed. This work proposes a stream-level adaptation of the current latency measures based on a re-segmentation approach applied to the output translation, that is successfully evaluated on streaming conditions for a reference IWSLT task + Simultaneous Machine Translation is the task of incrementally translating an input sentence before it is fully available. Currently, simultaneous translation is carried out by translating each sentence independently of the previously translated text. More generally, Streaming MT can be understood as an extension of Simultaneous MT to the incremental translation of a continuous input text stream. In this work, a state-of-the-art simultaneous sentence-level MT system is extended to the streaming setup by leveraging the streaming history. Extensive empirical results are reported on IWSLT Translation Tasks, showing that leveraging the streaming history leads to significant quality gains. In particular, the proposed system proves to compare favorably to the best performing systems. 2022.acl-long.480 2022.acl-long.480.software.zip iranzo-sanchez-etal-2022-simultaneous From 3803a03574dc18094c8f208151ae009e0c844b23 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 20:54:07 -0600 Subject: [PATCH 28/69] Paper pdf correction for 2023.emnlp-main.167, closes #2932. --- data/xml/2023.emnlp.xml | 1 + 1 file changed, 1 insertion(+) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 78118073c9..89ab722a37 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -2041,6 +2041,7 @@ lee-etal-2023-framework Fixed the sponsor in the Acknowledgments section. + Fixed the sponsor in the Acknowledgments section. Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair Modeling From 13d04be50192ba195bbf4dfa991f2ce9a5520aab Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 20:56:23 -0600 Subject: [PATCH 29/69] 2023.emnlp-main.167 fix. --- data/xml/2023.emnlp.xml | 1 - 1 file changed, 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 89ab722a37..78118073c9 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -2041,7 +2041,6 @@ lee-etal-2023-framework Fixed the sponsor in the Acknowledgments section. - Fixed the sponsor in the Acknowledgments section. Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair Modeling From 5fb689639c316a7359d0272e46101d348e937571 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 20:58:25 -0600 Subject: [PATCH 30/69] Paper pdf correction for 2023.jeptalnrecital-short.3, closes #2934. --- data/xml/2023.jeptalnrecital.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.jeptalnrecital.xml b/data/xml/2023.jeptalnrecital.xml index 7cd53adf60..56e12c1a35 100644 --- a/data/xml/2023.jeptalnrecital.xml +++ b/data/xml/2023.jeptalnrecital.xml @@ -364,9 +364,11 @@ MathieuConstant 23–36 Au début du XXIe siècle, le français faisait encore partie des langues peu dotées. Grâce aux efforts de la communauté française du traitement automatique des langues (TAL), de nombreuses ressources librement disponibles ont été produites, dont des lexiques du français. À travers cet article, nous nous intéressons à leur devenir dans la communauté par le prisme des actes de la conférence TALN sur une période de 20 ans. - 2023.jeptalnrecital-short.3 + 2023.jeptalnrecital-short.3 fra choi-etal-2023-des + + This version corrects a typo in the English abstract (ill-formed translation from the original abstract in French). Attention sur les spans pour l’analyse syntaxique en constituants From 74a2b6ce2824874e9ace96e4b59942d0851f33d4 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 20:59:43 -0600 Subject: [PATCH 31/69] Paper title meta data correction for 2023.findings-emnlp.231, closes #2936. --- data/xml/2023.findings.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 53800674be..653e0603ba 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -17110,7 +17110,7 @@ hou-smith-2023-detecting - A Word Sense Distribution-based approach for Semantic Change Prediction + Can Word Sense Distribution Detect Semantic Changes of Words? XiaohangTang YiZhou TaichiAida From c7b42cd01c9d9226d4be3d884f1a208323a210c2 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:02:09 -0600 Subject: [PATCH 32/69] Paper meta data correction for 2023.emnlp-main.900, closes #2938. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 78118073c9..966e8d148f 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -10790,7 +10790,7 @@ Bridging Continuous and Discrete Spaces: Interpretable Sentence Representation Learning via Compositional Operations - JamesHuang + James Y.Huang WenlinYao KaiqiangSong HongmingZhang From 2c56d85e85f13f46efbb7fd04dd51f96011d18f9 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:05:06 -0600 Subject: [PATCH 33/69] Author name meta data correction for Author Xuan Long Do, closes #2939. --- data/xml/2023.emnlp.xml | 2 +- data/xml/2023.findings.xml | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 966e8d148f..fe8e712995 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -10861,7 +10861,7 @@ <fixed-case>U</fixed-case>ni<fixed-case>C</fixed-case>hart: A Universal Vision-language Pretrained Model for Chart Comprehension and Reasoning AhmedMasry ParsaKavehzadeh - DoLong + Xuan LongDo EnamulHoque ShafiqJoty 14662-14684 diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 653e0603ba..2d027bea74 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -18103,7 +18103,7 @@ HailinChen WeishiWang FangkaiJiao - DoLong + Xuan LongDo ChengweiQin BoshengDing XiaobaoGuo From 2a94128e344e9c00c462047b654c0e4932f3c88a Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:08:32 -0600 Subject: [PATCH 34/69] Paper pdf correction for 2023.emnlp-main.771, closes #2940. --- data/xml/2023.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index fe8e712995..39119110e5 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -9249,8 +9249,10 @@ PijiLi 12522-12537 Open-domain multi-turn dialogue generation encounters the significant challenge of lacking various types of knowledge from diverse sources. Existing models typically focus on identifying specific types of dialogue knowledge and utilize corresponding datasets for training. However, this approach often leads to limited generalization capabilities and increased computational resource requirements. Recently, large language models (LLMs) have shown impressive performance on natural language processing tasks. To harness the knowledge storage of LLMs, we propose a framework named KnowEE that explores multi-source multi-type knowledge from LLMs by leveraging diverse datasets and then exploits the obtained knowledge for response generation. Our framework comprises two phases: First, we leverage five external datasets encompassing various types of knowledge to extract the most relevant samples to the dialogue context which are served as prompts to generate corresponding type of knowledge; Second, we inject the acquired knowledge into the ongoing dialogue context in fine-grained and coarse-grained manners, which is then fed into LLMs to generate the final dialogue response. Both automatic and manual evaluation results validate the effectiveness of our framework in exploring and exploiting multi-source multi-type knowledge to generate coherent, informative, and fluent responses. - 2023.emnlp-main.771 + 2023.emnlp-main.771 ni-etal-2023-multi + + Typo fix. Focus Your Attention (with Adaptive <fixed-case>IIR</fixed-case> Filters) From 212b0724e5c5a0f3611c727e12744237bd044bad Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:10:09 -0600 Subject: [PATCH 35/69] Volume meta data update for NLP-OSS 2023, closes #2941. --- data/xml/2023.nlposs.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.nlposs.xml b/data/xml/2023.nlposs.xml index 6e1efe7ea9..cc3ec18ea8 100644 --- a/data/xml/2023.nlposs.xml +++ b/data/xml/2023.nlposs.xml @@ -8,8 +8,8 @@ GeetickaChauhan JeremyGwinnup ElijahRippeth - Empirical Methods in Natural Language Processing -
Singapore, Singapore
+ Association for Computational Linguistics +
Singapore
December 2023 2023.nlposs-1 From 2e93deee8f7dcd48d02792eafc36a189d3d8cd9f Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:11:32 -0600 Subject: [PATCH 36/69] Paper pdf correction for 2023.emnlp-main.429, closes #2942. --- data/xml/2023.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 39119110e5..7708fbfa21 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -5186,8 +5186,10 @@ YuanjunLaili 6932-6953 Large Language Models (LLMs) have achieved remarkable success in many formal language oriented tasks, such as structural data-to-text and semantic parsing. However current benchmarks mostly follow the data distribution of the pre-training data of LLMs. Therefore, a natural question rises that do LLMs really understand the structured semantics of formal languages. In this paper, we investigate this problem on a special case, converse binary relation. We introduce a new benchmark ConvRe focusing on converse relations, which contains 17 relations and 1240 triples extracted from popular knowledge graph completion datasets. Our ConvRE features two tasks, Re2Text and Text2Re, which are formulated as multi-choice question answering to evaluate LLMs’ ability to determine the matching between relations and associated text. For the evaluation protocol, apart from different prompting methods, we further introduce variants to the test text and few-shot example text. We conduct experiments on three popular LLM families and have observed various scaling trends. The results suggest that LLMs often resort to shortcut learning and still face challenges on our proposed benchmark. - 2023.emnlp-main.429 + 2023.emnlp-main.429 qi-etal-2023-investigation + + This revision corrects the copying error in Table 7.
Towards Low-Resource Automatic Program Repair with Meta-Learning and Pretrained Language Models From 5fbaf6ab13951dce377d03ffae0e3b2dde1cb170 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:12:57 -0600 Subject: [PATCH 37/69] Paper pdf correction for 2023.findings-emnlp.949, closes #2949. --- data/xml/2023.findings.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 2d027bea74..ec4b5d38b1 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -25628,8 +25628,10 @@ AnnemarieFriedrich 14229-14241 Generative language models have recently shown remarkable success in generating answers to questions in a given textual context. However, these answers may suffer from hallucination, wrongly cite evidence, and spread misleading information. In this work, we address this problem by employing ChatGPT, a state-of-the-art generative model, as a machine-reading system. We ask it to retrieve answers to lexically varied and open-ended questions from trustworthy instructive texts. We introduce WHERE (WikiHow Evidence REtrieval), a new high-quality evaluation benchmark of a set of WikiHow articles exhaustively annotated with evidence sentences to questions that comes with a special challenge: All questions are about the article’s topic, but not all can be answered using the provided context. We interestingly find that when using a regular question-answering prompt, ChatGPT neglects to detect the unanswerable cases. When provided with a few examples, it learns to better judge whether a text provides answer evidence or not. Alongside this important finding, our dataset defines a new benchmark for evidence retrieval in question answering, which we argue is one of the necessary next steps for making large language models more trustworthy. - 2023.findings-emnlp.949 + 2023.findings-emnlp.949 henning-etal-2023-answer + + This revision provides a corrected version of Figure 1. <fixed-case>P</fixed-case>a<fixed-case>R</fixed-case>a<fixed-case>D</fixed-case>e: Passage Ranking using Demonstrations with <fixed-case>LLM</fixed-case>s From 0292c6309dee776f8818d5d87af13bd95d45d5a3 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:14:04 -0600 Subject: [PATCH 38/69] Paper pdf correction for 2023.emnlp-main.167, closes #2952. --- data/xml/2023.emnlp.xml | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 7708fbfa21..f25fc33243 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -2037,10 +2037,11 @@ HarksooKim 2789-2799 Most research on multimodal open-domain dialogue agents has focused on pretraining and multi-task learning using additional rich datasets beyond a given target dataset. However, methods for exploiting these additional datasets can be quite limited in real-world settings, creating a need for more efficient methods for constructing agents based solely on the target dataset. To address these issues, we present a new learning strategy called vision-language warm-up tasks for multimodal dialogue models (VLAW-MDM). This strategy does not require the use of large pretraining or multi-task datasets but rather relies solely on learning from target data. Moreover, our proposed approach automatically generate captions for images and incorporate them into the model’s input to improve the contextualization of visual information. Using this novel approach, we empirically demonstrate that our learning strategy is effective for limited data and relatively small models. The result show that our method achieved comparable and in some cases superior performance compared to existing state-of-the-art models on various evaluation metrics. - 2023.emnlp-main.167 + 2023.emnlp-main.167 lee-etal-2023-framework Fixed the sponsor in the Acknowledgments section. + Fixed the sponsor in the Acknowledgments section. Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair Modeling From 414516d7764ab7e47df6bff2eab99c69b3e791fa Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:15:17 -0600 Subject: [PATCH 39/69] Paper pdf correction for 2023.emnlp-main.633, closes #2953. --- data/xml/2023.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index f25fc33243..34fea90993 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -7624,8 +7624,10 @@ YuzhongQu 10241-10259 While question answering over knowledge bases (KBQA) has shown progress in addressing factoid questions, KBQA with numerical reasoning remains relatively unexplored. In this paper, we focus on the complex numerical reasoning in KBQA, and propose a new task, NR-KBQA, which necessitates the ability to perform both multi-hop reasoning and numerical reasoning. We also design a logic form in Python format called PyQL to represent the reasoning process of numerical reasoning questions. To facilitate the development of NR-KBQA, we present a large NR-KBQA dataset called MarkQA, which is automatically constructed by a small set of seeds. Each question in MarkQA is annotated with its corresponding SPARQL query, alongside the step-by-step reasoning path in the QDMR format and PyQL program. Experimental results of some state-of-the-art QA methods performed on the MarkQA dataset show that complex numerical reasoning in KBQA faces great challenges. - 2023.emnlp-main.633 + 2023.emnlp-main.633 huang-etal-2023-markqa + + Various fixes. Comparing Biases and the Impact of Multilingual Training across Multiple Languages From 1d079dcae0811903dedf02a675adf30d2344763d Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:16:28 -0600 Subject: [PATCH 40/69] Paper pdf correction for 2023.emnlp-main.814, closes #2957. --- data/xml/2023.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 34fea90993..a986537b7c 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -9744,8 +9744,10 @@ XipengQiu 13153-13187 Large language models (LLMs) can be used to serve as agents to simulate human behaviors, given the powerful ability to understand human instructions and provide high-quality generated texts. Such ability stimulates us to wonder whether LLMs can simulate a person in a higher form than simple human behaviors. Therefore, we aim to train an agent with the profile, experience, and emotional states of a specific person instead of using limited prompts to instruct ChatGPT API. In this work, we introduce Character-LLM that teach LLMs to act as specific people such as Beethoven, Queen Cleopatra, Julius Caesar, etc. Our method focuses on editing profiles as experiences of a certain character and training models to be personal simulacra with these experiences. To assess the effectiveness of our approach, we build a test playground that interviews trained agents and evaluates whether the agents memorize their characters and experiences. Experimental results show interesting observations that help build future simulacra of humankind. - 2023.emnlp-main.814 + 2023.emnlp-main.814 shao-etal-2023-character + + This revision corrects the footnote about the author on page 1. Natural Language Decompositions of Implicit Content Enable Better Text Representations From e87f463e299ca90ee01c6d6f9acc942dd2ded371 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Fri, 15 Dec 2023 21:17:39 -0600 Subject: [PATCH 41/69] Paper pdf correction for 2023.acl-long.519, closes #2958. --- data/xml/2023.acl.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.acl.xml b/data/xml/2023.acl.xml index 5498b6da30..0def38e001 100644 --- a/data/xml/2023.acl.xml +++ b/data/xml/2023.acl.xml @@ -6917,9 +6917,11 @@ YueZhangWestlake University 9332-9351 Most existing cross-lingual summarization (CLS) work constructs CLS corpora by simply and directly translating pre-annotated summaries from one language to another, which can contain errors from both summarization and translation processes. To address this issue, we propose ConvSumX, a cross-lingual conversation summarization benchmark, through a new annotation schema that explicitly considers source input context. ConvSumX consists of 2 sub-tasks under different real-world scenarios, with each covering 3 language directions. We conduct thorough analysis on ConvSumX and 3 widely-used manually annotated CLS corpora and empirically find that ConvSumX is more faithful towards input text. Additionally, based on the same intuition, we propose a 2-Step method, which takes both conversation and summary as input to simulate human annotation process. Experimental results show that 2-Step method surpasses strong baselines on ConvSumX under both automatic and human evaluation. Analysis shows that both source input text and summary are crucial for modeling cross-lingual summaries. - 2023.acl-long.519 + 2023.acl-long.519 chen-etal-2023-revisiting 10.18653/v1/2023.acl-long.519 + + Correct acknowledgement. Learning Dynamic Contextualised Word Embeddings via Template-based Temporal Adaptation From 4995dad835dcdaebf7add34a7a2a930f2e14d709 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:19:47 -0600 Subject: [PATCH 42/69] Paper pdf correction for 2023.findings-emnlp.279, closes #2969. --- data/xml/2023.findings.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index ec4b5d38b1..98316b54f5 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -17709,8 +17709,10 @@ LluísPadró 4234-4240 In this paper, we investigate the impact of objects on gender bias in image captioning systems. Our results show that only gender-specific objects have a strong gender bias (e.g., women-lipstick). In addition, we propose a visual semantic-based gender score that measures the degree of bias and can be used as a plug-in for any image captioning system. Our experiments demonstrate the utility of the gender score, since we observe that our score can measure the bias relation between a caption and its related gender; therefore, our score can be used as an additional metric to the existing Object Gender Co-Occ approach. - 2023.findings-emnlp.279 + 2023.findings-emnlp.279 sabir-padro-2023-women + + Fixed figure 1. <fixed-case>FREDS</fixed-case>um: A Dialogue Summarization Corpus for <fixed-case>F</fixed-case>rench Political Debates From 81ad4ad5c397db11f8273c9fd2475bd817e90678 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:21:47 -0600 Subject: [PATCH 43/69] Meta data correction for 2023.findings-emnlp.325, closes #2970. --- data/xml/2023.findings.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 98316b54f5..f1b7a7f49a 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -18240,7 +18240,7 @@ Controllable Chest <fixed-case>X</fixed-case>-Ray Report Generation from Longitudinal Representations - FrancescoSerra + FrancescoDalla Serra ChaoyangWang FaniDeligianni JeffDalton From a2e64ffa4afbb77ba6de91ae1e65068691b32a02 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:26:40 -0600 Subject: [PATCH 44/69] Author V.S.D.S.Mahesh Akavarapu meta data correction, closes #2967. --- data/xml/2023.wsc.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.wsc.xml b/data/xml/2023.wsc.xml index 4be1765d36..8706fec7c8 100644 --- a/data/xml/2023.wsc.xml +++ b/data/xml/2023.wsc.xml @@ -67,7 +67,7 @@ Creation of a Digital Rig <fixed-case>V</fixed-case>edic Index (Anukramani) for Computational Linguistic Tasks - A V S D SMahesh + V.S.D.S.MaheshAkavarapu ArnabBhattacharya 89–96 2023.wsc-csdh.6 From e972b16eb14b091fa8f49c94357ac7fbc2244be5 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:44:01 -0600 Subject: [PATCH 45/69] Workshop meta data correction for 2023.bigpicture-1 & 2023.newsum-1, closes #2965. --- data/xml/2023.bigpicture.xml | 2 +- data/xml/2023.newsum.xml | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.bigpicture.xml b/data/xml/2023.bigpicture.xml index 63890542e6..4bd89588de 100644 --- a/data/xml/2023.bigpicture.xml +++ b/data/xml/2023.bigpicture.xml @@ -9,7 +9,7 @@ SebastianRuder NoahA. Smith Association for Computational Linguistics -
Singapore, Singapore
+
Singapore
December 2023 2023.bigpicture-1 diff --git a/data/xml/2023.newsum.xml b/data/xml/2023.newsum.xml index bfd364c17c..0e341341e0 100644 --- a/data/xml/2023.newsum.xml +++ b/data/xml/2023.newsum.xml @@ -9,7 +9,7 @@ FeiLiu GiuseppeCarenini Association for Computational Linguistics -
Hybrid
+
Singapore
December 2023 2023.newsum-1 From e2e7dd8259beb076ea1f86699e9ea131f02935df Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:48:02 -0600 Subject: [PATCH 46/69] Meta data correction for 2023.emnlp-main.166, closes #2964. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index a986537b7c..e2dcdbdeaa 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -2021,7 +2021,7 @@ Indicative Summarization of Long Discussions ShahbazSyed DominikSchwabe - KhalidKhatib + KhalidAl-Khatib MartinPotthast 2752-2788 Online forums encourage the exchange and discussion of different stances on many topics. Not only do they provide an opportunity to present one’s own arguments, but may also gather a broad cross-section of others’ arguments. However, the resulting long discussions are difficult to overview. This paper presents a novel unsupervised approach using large language models (LLMs) to generating indicative summaries for long discussions that basically serve as tables of contents. Our approach first clusters argument sentences, generates cluster labels as abstractive summaries, and classifies the generated cluster labels into argumentation frames resulting in a two-level summary. Based on an extensively optimized prompt engineering approach, we evaluate 19 LLMs for generative cluster labeling and frame classification. To evaluate the usefulness of our indicative summaries, we conduct a purpose-driven user study via a new visual interface called **Discussion Explorer**: It shows that our proposed indicative summaries serve as a convenient navigation tool to explore long discussions. From 806c3e69459602d8266f97155b10227056c8e335 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:54:25 -0600 Subject: [PATCH 47/69] Paper pdf correction for 2023.emnlp-main.73, closes #2961. --- data/xml/2023.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index e2dcdbdeaa..5579930f3d 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -886,8 +886,10 @@ XipengQiu 1144-1156 Widely applied large language models (LLMs) can generate human-like content, raising concerns about the abuse of LLMs. Therefore, it is important to build strong AI-generated text (AIGT) detectors. Current works only consider document-level AIGT detection, therefore, in this paper, we first introduce a sentence-level detection challenge by synthesizing a dataset that contains documents that are polished with LLMs, that is, the documents contain sentences written by humans and sentences modified by LLMs. Then we propose Sequence X (Check) GPT, a novel method that utilizes log probability lists from white-box LLMs as features for sentence-level AIGT detection. These features are composed like waves in speech processing and cannot be studied by LLMs. Therefore, we build SeqXGPT based on convolution and self-attention networks. We test it in both sentence and document-level detection challenges. Experimental results show that previous methods struggle in solving sentence-level AIGT detection, while our method not only significantly surpasses baseline methods in both sentence and document-level detection challenges but also exhibits strong generalization capabilities. - 2023.emnlp-main.73 + 2023.emnlp-main.73 wang-etal-2023-seqxgpt + + Fixed footnote on page 1.
<fixed-case>QTS</fixed-case>umm: Query-Focused Summarization over Tabular Data From 564c1710f9058357684964c31023c0d0c98a3ad4 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:56:04 -0600 Subject: [PATCH 48/69] Paper meta data correction for 2023.arabicnlp-1.55, closes #2945. --- data/xml/2023.arabicnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.arabicnlp.xml b/data/xml/2023.arabicnlp.xml index 64d9435483..7e2970e6be 100644 --- a/data/xml/2023.arabicnlp.xml +++ b/data/xml/2023.arabicnlp.xml @@ -628,7 +628,7 @@ <fixed-case>UL</fixed-case> & <fixed-case>UM</fixed-case>6<fixed-case>P</fixed-case> at <fixed-case>A</fixed-case>r<fixed-case>AIE</fixed-case>val Shared Task: Transformer-based model for Persuasion Techniques and Disinformation detection in <fixed-case>A</fixed-case>rabic SalimaLamsiyah - AbdelkaderMahdaouy + AbdelkaderEl Mahdaouy HamzaAlami IsmailBerrada ChristophSchommer From 23b1ae943e949b9ed92083d8f5d49f797b79fc35 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 13:57:24 -0600 Subject: [PATCH 49/69] Paper meta data correction for 2023.arabicnlp-1.87, closes #2946. --- data/xml/2023.arabicnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.arabicnlp.xml b/data/xml/2023.arabicnlp.xml index 7e2970e6be..98fef5ab71 100644 --- a/data/xml/2023.arabicnlp.xml +++ b/data/xml/2023.arabicnlp.xml @@ -964,7 +964,7 @@ <fixed-case>UM</fixed-case>6<fixed-case>P</fixed-case> & <fixed-case>UL</fixed-case> at <fixed-case>W</fixed-case>ojood<fixed-case>NER</fixed-case> shared task: Improving Multi-Task Learning for Flat and Nested <fixed-case>A</fixed-case>rabic Named Entity Recognition - AbdelkaderMahdaouy + AbdelkaderEl Mahdaouy SalimaLamsiyah HamzaAlami ChristophSchommer From cd80051b60c49ca8cf37954adc2e39234b451cd2 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 15:34:27 -0600 Subject: [PATCH 50/69] Author meta data correction for 2023.genbench-1.4, closes #2960. --- data/xml/2023.genbench.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.genbench.xml b/data/xml/2023.genbench.xml index 09bf552ab6..b1f42cde74 100644 --- a/data/xml/2023.genbench.xml +++ b/data/xml/2023.genbench.xml @@ -62,7 +62,7 @@ Evaluating Neural Language Models as Cognitive Models of Language Acquisition - HectorJavier Vazquez MartinezUniversity of Pennsylvania + HéctorVázquez MartínezUniversity of Pennsylvania AnnikaLea HeuserUniversity of Pennsylvania CharlesYangUniversity of Pennsylvania JordanKodnerStony Brook University From 7edfa2b2c0c3c214e56fc869271d0ffe6ad6ac24 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 15:38:24 -0600 Subject: [PATCH 51/69] Paper meta data correction for 2023.findings-emnlp.993, closes #2959. --- data/xml/2023.findings.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index f1b7a7f49a..8cda03c881 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -26140,7 +26140,7 @@ Qualitative Code Suggestion: A Human-Centric Approach to Qualitative Coding - CesarePiano + CesareSpinoso-Di Piano SamiraRahimi JackieCheung 14887-14909 From c0e9f9f2d7ee8607cd0d1346ba659ff9af17d7f4 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 15:53:38 -0600 Subject: [PATCH 52/69] Author meta data correction for 2023.findings-emnlp.469, closes #2956. --- data/xml/2023.findings.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 8cda03c881..04fdcdb57d 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -19939,7 +19939,7 @@ Visually Grounded Continual Language Learning with Selective Specialization KyraAhrens LennartBengtson - JaeLee + JaeHee Lee StefanWermter 7037-7054 A desirable trait of an artificial agent acting in the visual world is to continually learn a sequence of language-informed tasks while striking a balance between sufficiently specializing in each task and building a generalized knowledge for transfer. Selective specialization, i.e., a careful selection of model components to specialize in each task, is a strategy to provide control over this trade-off. However, the design of selection strategies requires insights on the role of each model component in learning rather specialized or generalizable representations, which poses a gap in current research. Thus, our aim with this work is to provide an extensive analysis of selection strategies for visually grounded continual language learning. Due to the lack of suitable benchmarks for this purpose, we introduce two novel diagnostic datasets that provide enough control and flexibility for a thorough model analysis. We assess various heuristics for module specialization strategies as well as quantifiable measures for two different types of model architectures. Finally, we design conceptually simple approaches based on our analysis that outperform common continual learning baselines. Our results demonstrate the need for further efforts towards better aligning continual learning algorithms with the learning behaviors of individual model parts. From 7b147577a54bd520c974057048854a1444203fc3 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 15:55:54 -0600 Subject: [PATCH 53/69] Author meta data correction for 2023.emnlp-main.783, closes #2951. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 5579930f3d..fb3425ed8e 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -9383,7 +9383,7 @@ Explaining Interactions Between Text Spans - SagnikChoudhury + SagnikRay Choudhury PepaAtanasova IsabelleAugenstein 12709-12730 From 5140999668f921a19f1e6b8d544e78566c16937e Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 15:58:01 -0600 Subject: [PATCH 54/69] Author meta data correction for 2023.emnlp-main.953, closes #2947. --- data/xml/2023.emnlp.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index fb3425ed8e..05f5d9e03e 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -11465,11 +11465,11 @@ Automatic Transcription of Handwritten Old <fixed-case>O</fixed-case>ccitan Language - EstebanArias + EstebanGarces Arias VallariPai MatthiasSchöffel ChristianHeumann - MatthiasAenmacher + MatthiasAßenmacher 15416-15439 While existing neural network-based approaches have shown promising results in Handwritten Text Recognition (HTR) for high-resource languages and standardized/machine-written text, their application to low-resource languages often presents challenges, resulting in reduced effectiveness. In this paper, we propose an innovative HTR approach that leverages the Transformer architecture for recognizing handwritten Old Occitan language. Given the limited availability of data, which comprises only word pairs of graphical variants and lemmas, we develop and rely on elaborate data augmentation techniques for both text and image data. Our model combines a custom-trained Swin image encoder with a BERT text decoder, which we pre-train using a large-scale augmented synthetic data set and fine-tune on the small human-labeled data set. Experimental results reveal that our approach surpasses the performance of current state-of-the-art models for Old Occitan HTR, including open-source Transformer-based models such as a fine-tuned TrOCR and commercial applications like Google Cloud Vision. To nurture further research and development, we make our models, data sets, and code publicly available. 2023.emnlp-main.953 From 96f7dc321a9f3a363edcaf0a3eb1633b2738695c Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Tue, 19 Dec 2023 19:54:46 -0600 Subject: [PATCH 55/69] Author meta data correction for 2023.emnlp-main.470, closes #2972. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 05f5d9e03e..bed64db09d 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -5663,7 +5663,7 @@ <fixed-case>W</fixed-case>i<fixed-case>CE</fixed-case>: Real-World Entailment for Claims in <fixed-case>W</fixed-case>ikipedia RyoKamoi TanyaGoyal - JuanRodriguez + JuanDiego Rodriguez GregDurrett 7561-7583 Textual entailment models are increasingly applied in settings like fact-checking, presupposition verification in question answering, or summary evaluation. However, these represent a significant domain shift from existing entailment datasets, and models underperform as a result. We propose WiCE, a new fine-grained textual entailment dataset built on natural claim and evidence pairs extracted from Wikipedia. In addition to standard claim-level entailment, WiCE provides entailment judgments over sub-sentence units of the claim, and a minimal subset of evidence sentences that support each subclaim. To support this, we propose an automatic claim decomposition strategy using GPT-3.5 which we show is also effective at improving entailment models’ performance on multiple datasets at test time. Finally, we show that real claims in our dataset involve challenging verification and retrieval problems that existing models fail to address. From a6eb9147a3f72c55becfe66cbd4ea4dc1e621846 Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Thu, 21 Dec 2023 19:58:33 -0600 Subject: [PATCH 56/69] Paper meta data correction for 2023.emnlp-main.237, closes #2978. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index bed64db09d..f16a5f7571 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -2882,7 +2882,7 @@ <fixed-case>COVID</fixed-case>-19 Vaccine Misinformation in Middle Income Countries JonginKim - ByeoBak + Byeo RheeBak AdityaAgrawal JiaxiWu VeronikaWirtz From 631d694c4a81c1c6b14c63b58e22a78db3bca91e Mon Sep 17 00:00:00 2001 From: anthology-assist Date: Thu, 21 Dec 2023 20:00:40 -0600 Subject: [PATCH 57/69] Author name meta data correction for 2023.arabicnlp-1.53, closes #2980. --- data/xml/2023.arabicnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.arabicnlp.xml b/data/xml/2023.arabicnlp.xml index 98fef5ab71..cedfd84486 100644 --- a/data/xml/2023.arabicnlp.xml +++ b/data/xml/2023.arabicnlp.xml @@ -610,7 +610,7 @@ Itri Amigos at <fixed-case>A</fixed-case>r<fixed-case>AIE</fixed-case>val Shared Task: Transformer vs. Compression-Based Models for Persuasion Techniques and Disinformation Detection JehadOumer NoumanAhmed - NataliaManrique + NataliaFlechas Manrique 543-548 Social media has significantly amplified the dissemination of misinformation. Researchers have employed natural language processing and machine learning techniques to identify and categorize false information on these platforms. While there is a well-established body of research on detecting fake news in English and Latin languages, the study of Arabic fake news detection remains limited. This paper describes the methods used to tackle the challenges of the ArAIEval shared Task 2023. We conducted experiments with both monolingual Arabic and multi-lingual pre-trained Language Models (LM). We found that the monolingual Arabic models outperformed in all four subtasks. Additionally, we explored a novel lossless compression method, which, while not surpassing pretrained LM performance, presents an intriguing avenue for future experimentation to achieve comparable results in a more efficient and rapid manner. 2023.arabicnlp-1.53 From 79aef796bbe357db4205a4436f6ebb70db10e29a Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 13:49:00 -0600 Subject: [PATCH 58/69] Paper pdf correction for 2023.acl-long.513, closes #2982. --- data/xml/2023.acl.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.acl.xml b/data/xml/2023.acl.xml index 0def38e001..ac8823844b 100644 --- a/data/xml/2023.acl.xml +++ b/data/xml/2023.acl.xml @@ -6827,9 +6827,11 @@ NanyunPengUniversity of California, Los Angeles 9235-9254 Automatic melody-to-lyric generation is a task in which song lyrics are generated to go with a given melody. It is of significant practical interest and more challenging than unconstrained lyric generation as the music imposes additional constraints onto the lyrics. The training data is limited as most songs are copyrighted, resulting in models that underfit the complicated cross-modal relationship between melody and lyrics. In this work, we propose a method for generating high-quality lyrics without training on any aligned melody-lyric data. Specifically, we design a hierarchical lyric generation framework that first generates a song outline and second the complete lyrics. The framework enables disentanglement of training (based purely on text) from inference (melody-guided text generation) to circumvent the shortage of parallel data. We leverage the segmentation and rhythm alignment between melody and lyrics to compile the given melody into decoding constraints as guidance during inference. The two-step hierarchical design also enables content control via the lyric outline, a much-desired feature for democratizing collaborative song creation. Experimental results show that our model can generate high-quality lyrics that are more on-topic, singable, intelligible, and coherent than strong baselines, for example SongMASS, a SOTA model trained on a parallel dataset, with a 24% relative overall quality improvement based on human ratings. Our code is available at https://github.com/amazon-science/unsupervised-melody-to-lyrics-generation. - 2023.acl-long.513 + 2023.acl-long.513 tian-etal-2023-unsupervised 10.18653/v1/2023.acl-long.513 + + Added description of authors contributions. Causality-aware Concept Extraction based on Knowledge-guided Prompting From d0b518fee5bd524715eae6d80513c483740c2511 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 13:50:15 -0600 Subject: [PATCH 59/69] Paper pdf correction for 2023.emnlp-main.434, closes #2985. --- data/xml/2023.emnlp.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 2b4fcc7626..24298318b8 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -5679,9 +5679,11 @@ RyanCotterell 7011-7034 This work investigates the computational expressivity of language models (LMs) based on recurrent neural networks (RNNs). Siegelmann and Sontag (1992) famously showed that RNNs with rational weights and hidden states and unbounded computation time are Turing complete. However, LMs define weightings over strings in addition to just (unweighted) language membership and the analysis of the computational power of RNN LMs (RLMs) should reflect this. We extend the Turing completeness result to the probabilistic case, showing how a rationally weighted RLM with unbounded computation time can simulate any deterministic probabilistic Turing machine (PTM) with rationally weighted transitions. Since, in practice, RLMs work in real-time, processing a symbol at every time step, we treat the above result as an upper bound on the expressivity of RLMs. We also provide a lower bound by showing that under the restriction to real-time computation, such models can simulate deterministic real-time rational PTMs. - 2023.emnlp-main.434 + 2023.emnlp-main.434 nowak-etal-2023-representational 10.18653/v1/2023.emnlp-main.434 + + Required that the weighting functions in definition are non-negative. A Mechanistic Interpretation of Arithmetic Reasoning in Language Models using Causal Mediation Analysis From d31df5970196b750e45942b6fe83b96c7abd0cf1 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 13:51:30 -0600 Subject: [PATCH 60/69] Paper pdf correction for 2023.findings-emnlp.709, closes #2991. --- data/xml/2023.findings.xml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index fcdcc6256c..26670d209b 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -23442,9 +23442,11 @@ YulanHe 10559-10571 Adjusting for latent covariates is crucial for estimating causal effects from observational textual data. Most existing methods only account for confounding covariates that affect both treatment and outcome, potentially leading to biased causal effects. This bias arises from insufficient consideration of non-confounding covariates, which are relevant only to either the treatment or the outcome. In this work, we aim to mitigate the bias by unveiling interactions between different variables to disentangle the non-confounding covariates when estimating causal effects from text. The disentangling process ensures covariates only contribute to their respective objectives, enabling independence between variables. Additionally, we impose a constraint to balance representations from the treated group and control group to alleviate selection bias. We conduct experiments on two different treatment factors under various scenarios, and the proposed model significantly outperforms recent strong baselines. Furthermore, our thorough analysis on earnings call transcripts demonstrates that our model can effectively disentangle the variables, and further investigations into real-world scenarios provide guidance for investors to make informed decisions. - 2023.findings-emnlp.709 + 2023.findings-emnlp.709 zhou-he-2023-causal 10.18653/v1/2023.findings-emnlp.709 + + Corrects the tick mark typo in Table 2. Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples! From 8ca9d100c5fb6519dd844868c84b0fb8c86f0219 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 13:56:14 -0600 Subject: [PATCH 61/69] Author meta data correctioon for 2023.findings-emnlp.382, closes #2984. --- data/xml/2023.findings.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index 26670d209b..e593deba74 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -19301,7 +19301,7 @@ Representativeness as a Forgotten Lesson for Multilingual and Code-switched Data Collection and Preparation A. SezaDoğruöz SunayanaSitaram - ZhengYong + Zheng XinYong 5751-5767 Multilingualism is widespread around the world and code-switching (CSW) is a common practice among different language pairs/tuples across locations and regions. However, there is still not much progress in building successful CSW systems, despite the recent advances in Massive Multilingual Language Models (MMLMs). We investigate the reasons behind this setback through a critical study about the existing CSW data sets (68) across language pairs in terms of the collection and preparation (e.g. transcription and annotation) stages. This in-depth analysis reveals that a) most CSW data involves English ignoring other language pairs/tuples b) there are flaws in terms of representativeness in data collection and preparation stages due to ignoring the location based, socio-demographic and register variation in CSW. In addition, lack of clarity on the data selection and filtering stages shadow the representativeness of CSW data sets. We conclude by providing a short check-list to improve the representativeness for forthcoming studies involving CSW data collection and preparation. 2023.findings-emnlp.382 From d7687fdb52ef7143cf2f951c63a9921bb83979c3 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 13:58:58 -0600 Subject: [PATCH 62/69] Paper meta data correction for 2023.emnlp-main.608, closes #2987. --- data/xml/2023.emnlp.xml | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 24298318b8..28a29c3892 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -7918,14 +7918,14 @@ 10.18653/v1/2023.emnlp-main.607 - A Video Is Worth 4096 Tokens: Verbalize Story Videos To Understand Them In Zero Shot + A Video Is Worth 4096 Tokens: Verbalize Videos To Understand Them In Zero Shot AanishaBhattacharyya - YamanSingla + Yaman KSingla BalajiKrishnamurthy - RajivShah + Rajiv RatnShah ChangyouChen 9822-9839 - Multimedia content, such as advertisements and story videos, exhibit a rich blend of creativity and multiple modalities. They incorporate elements like text, visuals, audio, and storytelling techniques, employing devices like emotions, symbolism, and slogans to convey meaning. There is a dearth of large annotated training datasets in the multimedia domain hindering the development of supervised learn-ing models with satisfactory performance for real-world applications. On the other hand, the rise of large language models (LLMs) has witnessed remarkable zero-shot performance in various natural language processing (NLP) tasks, such as emotion classification, question-answering, and topic classification. To leverage such advanced techniques to bridge this performance gap in multimedia understanding, we propose verbalizing long videos to generate their descriptions in natural language, followed by performing video-understanding tasks on the generated story as opposed to the original video. Through extensive experiments on fifteen video-understanding tasks, we demonstrate that our method, despite being zero-shot, achieves significantly better results than supervised baselines for video understanding. Furthermore, to alleviate a lack of story understanding benchmarks, we publicly release the first dataset on a crucial task in computational social science on persuasion strategy identification. + Multimedia content, such as advertisements and story videos, exhibit a rich blend of creativity and multiple modalities. They incorporate elements like text, visuals, audio, and storytelling techniques, employing devices like emotions, symbolism, and slogans to convey meaning. There is a dearth of large annotated training datasets in the multimedia domain hindering the development of supervised learning models with satisfactory performance for real-world applications. On the other hand, the rise of large language models (LLMs) has witnessed remarkable zero-shot performance in various natural language processing (NLP) tasks, such as emotion classification, question answering, and topic classification. To leverage such advanced techniques to bridge this performance gap in multimedia understanding, we propose verbalizing long videos to generate their descriptions in natural language, followed by performing video-understanding tasks on the generated story as opposed to the original video. Through extensive experiments on fifteen video-understanding tasks, we demonstrate that our method, despite being zero-shot, achieves significantly better results than supervised baselines for video understanding. Furthermore, to alleviate a lack of story understanding benchmarks, we publicly release the first dataset on a crucial task in computational social science on persuasion strategy identification. 2023.emnlp-main.608 bhattacharyya-etal-2023-video 10.18653/v1/2023.emnlp-main.608 From 7b3bd49339e7b27cf7c1d4ce1ae65810d5e25bca Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 14:01:19 -0600 Subject: [PATCH 63/69] Author meta data correction for 2023.emnlp-main.714, closes #2988. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 28a29c3892..266f22ece0 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -9306,7 +9306,7 @@ Generating Summaries with Controllable Readability Levels - LeonardoRibeiro + Leonardo F. R.Ribeiro MohitBansal MarkusDreyer 11669-11687 From 02a5748ee639b45c57f7fed11d8895fd85db5c08 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Sun, 31 Dec 2023 14:03:11 -0600 Subject: [PATCH 64/69] Author meta data correction for 2023.emnlp-main.724, closes #2990. --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 266f22ece0..6504a707c0 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -9434,7 +9434,7 @@ Back Transcription as a Method for Evaluating Robustness of Natural Language Understanding Models to Speech Recognition Errors MarekKubis PawełSkórzewski - MarcinSowańnski + MarcinSowański TomaszZietkiewicz 11824-11835 In a spoken dialogue system, an NLU model is preceded by a speech recognition system that can deteriorate the performance of natural language understanding. This paper proposes a method for investigating the impact of speech recognition errors on the performance of natural language understanding models. The proposed method combines the back transcription procedure with a fine-grained technique for categorizing the errors that affect the performance of NLU models. The method relies on the usage of synthesized speech for NLU evaluation. We show that the use of synthesized speech in place of audio recording does not change the outcomes of the presented technique in a significant way. From 040c0575fb6f16a3a5a972a15cf9fadb423356a5 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Thu, 4 Jan 2024 18:56:40 -0600 Subject: [PATCH 65/69] inserted license comment. --- python/acl_anthology/__init__.py | 2 +- python/acl_anthology/anthology.py | 2 +- python/acl_anthology/collections/__init__.py | 2 +- python/acl_anthology/collections/collection.py | 2 +- python/acl_anthology/collections/event.py | 2 +- python/acl_anthology/collections/eventindex.py | 2 +- python/acl_anthology/collections/index.py | 2 +- python/acl_anthology/collections/paper.py | 2 +- python/acl_anthology/collections/types.py | 2 +- python/acl_anthology/collections/volume.py | 2 +- python/acl_anthology/config.py | 2 +- python/acl_anthology/constants.py | 2 +- python/acl_anthology/containers.py | 2 +- python/acl_anthology/exceptions.py | 2 +- python/acl_anthology/files.py | 2 +- python/acl_anthology/people/__init__.py | 2 +- python/acl_anthology/people/index.py | 2 +- python/acl_anthology/people/name.py | 2 +- python/acl_anthology/people/person.py | 2 +- python/acl_anthology/sigs.py | 2 +- python/acl_anthology/text/__init__.py | 2 +- python/acl_anthology/text/markuptext.py | 2 +- python/acl_anthology/text/texmath.py | 2 +- python/acl_anthology/utils/__init__.py | 2 +- python/acl_anthology/utils/git.py | 2 +- python/acl_anthology/utils/ids.py | 2 +- python/acl_anthology/utils/latex.py | 2 +- python/acl_anthology/utils/logging.py | 2 +- python/acl_anthology/utils/text.py | 2 +- python/acl_anthology/utils/xml.py | 2 +- python/acl_anthology/venues.py | 2 +- python/benchmarks/bench_attrs.py | 2 +- python/benchmarks/bench_sanitycheck.py | 2 +- python/benchmarks/bench_utils.py | 2 +- python/benchmarks/bench_xml_markup.py | 2 +- python/benchmarks/bench_xml_names.py | 2 +- python/benchmarks/bench_xml_parsing.py | 2 +- python/tests/anthology_integration_test.py | 2 +- python/tests/anthology_test.py | 2 +- python/tests/collections/collection_test.py | 2 +- python/tests/collections/collectionindex_test.py | 2 +- python/tests/collections/event_test.py | 2 +- python/tests/collections/eventindex_test.py | 2 +- python/tests/collections/paper_test.py | 2 +- python/tests/collections/volume_test.py | 2 +- python/tests/conftest.py | 2 +- python/tests/containers_test.py | 2 +- python/tests/files_test.py | 2 +- python/tests/people/name_test.py | 2 +- python/tests/people/person_test.py | 2 +- python/tests/people/personindex_test.py | 2 +- python/tests/sigs_test.py | 2 +- python/tests/text/markuptext_test.py | 2 +- python/tests/text/texmath_test.py | 2 +- python/tests/utils/ids_test.py | 2 +- python/tests/utils/latex_test.py | 2 +- python/tests/utils/logging_test.py | 2 +- python/tests/utils/text_test.py | 2 +- python/tests/utils/xml_test.py | 2 +- python/tests/venues_test.py | 2 +- 60 files changed, 60 insertions(+), 60 deletions(-) diff --git a/python/acl_anthology/__init__.py b/python/acl_anthology/__init__.py index 382bf090d8..5d1c5dd731 100644 --- a/python/acl_anthology/__init__.py +++ b/python/acl_anthology/__init__.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/anthology.py b/python/acl_anthology/anthology.py index ebacdc3602..0761e017c7 100644 --- a/python/acl_anthology/anthology.py +++ b/python/acl_anthology/anthology.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/__init__.py b/python/acl_anthology/collections/__init__.py index df003b3bc1..a8d8b30f21 100644 --- a/python/acl_anthology/collections/__init__.py +++ b/python/acl_anthology/collections/__init__.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/collection.py b/python/acl_anthology/collections/collection.py index d99dda23b5..daa7c378ca 100644 --- a/python/acl_anthology/collections/collection.py +++ b/python/acl_anthology/collections/collection.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/event.py b/python/acl_anthology/collections/event.py index 750bc8f747..b6c7ae1fc5 100644 --- a/python/acl_anthology/collections/event.py +++ b/python/acl_anthology/collections/event.py @@ -1,5 +1,5 @@ # Copyright 2022 Matt Post -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/eventindex.py b/python/acl_anthology/collections/eventindex.py index ca701f9a88..43140dfd28 100644 --- a/python/acl_anthology/collections/eventindex.py +++ b/python/acl_anthology/collections/eventindex.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/index.py b/python/acl_anthology/collections/index.py index 1bf257523e..85904f2c97 100644 --- a/python/acl_anthology/collections/index.py +++ b/python/acl_anthology/collections/index.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/paper.py b/python/acl_anthology/collections/paper.py index 4ac25483b0..acbfb3448b 100644 --- a/python/acl_anthology/collections/paper.py +++ b/python/acl_anthology/collections/paper.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/types.py b/python/acl_anthology/collections/types.py index ef26af6639..e7ff137832 100644 --- a/python/acl_anthology/collections/types.py +++ b/python/acl_anthology/collections/types.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/collections/volume.py b/python/acl_anthology/collections/volume.py index c2d65c0928..43735fb646 100644 --- a/python/acl_anthology/collections/volume.py +++ b/python/acl_anthology/collections/volume.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/config.py b/python/acl_anthology/config.py index 4d9f0cbb0c..5d371e05e8 100644 --- a/python/acl_anthology/config.py +++ b/python/acl_anthology/config.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/constants.py b/python/acl_anthology/constants.py index 42d0800088..67fe017a16 100644 --- a/python/acl_anthology/constants.py +++ b/python/acl_anthology/constants.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/containers.py b/python/acl_anthology/containers.py index 58d1b3e92a..98a67decdd 100644 --- a/python/acl_anthology/containers.py +++ b/python/acl_anthology/containers.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/exceptions.py b/python/acl_anthology/exceptions.py index 0d16951c66..74830b36ef 100644 --- a/python/acl_anthology/exceptions.py +++ b/python/acl_anthology/exceptions.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/files.py b/python/acl_anthology/files.py index 0347356f9b..f8fded4fcc 100644 --- a/python/acl_anthology/files.py +++ b/python/acl_anthology/files.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/people/__init__.py b/python/acl_anthology/people/__init__.py index ba7e72f628..49063fa6de 100644 --- a/python/acl_anthology/people/__init__.py +++ b/python/acl_anthology/people/__init__.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/people/index.py b/python/acl_anthology/people/index.py index fb14469bc3..41b25cb993 100644 --- a/python/acl_anthology/people/index.py +++ b/python/acl_anthology/people/index.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/people/name.py b/python/acl_anthology/people/name.py index 4876dd0326..116dd17240 100644 --- a/python/acl_anthology/people/name.py +++ b/python/acl_anthology/people/name.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/people/person.py b/python/acl_anthology/people/person.py index 1f123985eb..075086c476 100644 --- a/python/acl_anthology/people/person.py +++ b/python/acl_anthology/people/person.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/sigs.py b/python/acl_anthology/sigs.py index 3475ca812b..9a835d2d43 100644 --- a/python/acl_anthology/sigs.py +++ b/python/acl_anthology/sigs.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/text/__init__.py b/python/acl_anthology/text/__init__.py index 6aa2046421..9622479a9d 100644 --- a/python/acl_anthology/text/__init__.py +++ b/python/acl_anthology/text/__init__.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/text/markuptext.py b/python/acl_anthology/text/markuptext.py index 2d81f67fd4..facc0e1c2e 100644 --- a/python/acl_anthology/text/markuptext.py +++ b/python/acl_anthology/text/markuptext.py @@ -1,4 +1,4 @@ -# Copyright 2019-2023 Marcel Bollmann +# Copyright 2019-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/text/texmath.py b/python/acl_anthology/text/texmath.py index 2dee8fbec1..4af8bb5cb9 100644 --- a/python/acl_anthology/text/texmath.py +++ b/python/acl_anthology/text/texmath.py @@ -1,4 +1,4 @@ -# Copyright 2019-2023 Marcel Bollmann +# Copyright 2019-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/__init__.py b/python/acl_anthology/utils/__init__.py index 5d4173b990..fcf9e8520d 100644 --- a/python/acl_anthology/utils/__init__.py +++ b/python/acl_anthology/utils/__init__.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/git.py b/python/acl_anthology/utils/git.py index f8891ee2c1..f203c75eca 100644 --- a/python/acl_anthology/utils/git.py +++ b/python/acl_anthology/utils/git.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/ids.py b/python/acl_anthology/utils/ids.py index 6e4434fe04..b70519e9a5 100644 --- a/python/acl_anthology/utils/ids.py +++ b/python/acl_anthology/utils/ids.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/latex.py b/python/acl_anthology/utils/latex.py index b1821357d7..a3c492f4de 100644 --- a/python/acl_anthology/utils/latex.py +++ b/python/acl_anthology/utils/latex.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/logging.py b/python/acl_anthology/utils/logging.py index e6119ce213..ae1fabaabe 100644 --- a/python/acl_anthology/utils/logging.py +++ b/python/acl_anthology/utils/logging.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/text.py b/python/acl_anthology/utils/text.py index d030a44a11..c4abda31ac 100644 --- a/python/acl_anthology/utils/text.py +++ b/python/acl_anthology/utils/text.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/utils/xml.py b/python/acl_anthology/utils/xml.py index 87942cb7da..f52f512afd 100644 --- a/python/acl_anthology/utils/xml.py +++ b/python/acl_anthology/utils/xml.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/acl_anthology/venues.py b/python/acl_anthology/venues.py index 16b05e822a..6043b4f465 100644 --- a/python/acl_anthology/venues.py +++ b/python/acl_anthology/venues.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/benchmarks/bench_attrs.py b/python/benchmarks/bench_attrs.py index dd2c4121a4..194224afce 100644 --- a/python/benchmarks/bench_attrs.py +++ b/python/benchmarks/bench_attrs.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/benchmarks/bench_sanitycheck.py b/python/benchmarks/bench_sanitycheck.py index 1659bc51f6..a22d91fc9e 100644 --- a/python/benchmarks/bench_sanitycheck.py +++ b/python/benchmarks/bench_sanitycheck.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/benchmarks/bench_utils.py b/python/benchmarks/bench_utils.py index 1953223505..e7a69861fe 100644 --- a/python/benchmarks/bench_utils.py +++ b/python/benchmarks/bench_utils.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/benchmarks/bench_xml_markup.py b/python/benchmarks/bench_xml_markup.py index a3b4873667..47185ff074 100644 --- a/python/benchmarks/bench_xml_markup.py +++ b/python/benchmarks/bench_xml_markup.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/benchmarks/bench_xml_names.py b/python/benchmarks/bench_xml_names.py index 401f94fbe1..d729eba94c 100644 --- a/python/benchmarks/bench_xml_names.py +++ b/python/benchmarks/bench_xml_names.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/benchmarks/bench_xml_parsing.py b/python/benchmarks/bench_xml_parsing.py index 219a425bfb..c1c9b3a697 100644 --- a/python/benchmarks/bench_xml_parsing.py +++ b/python/benchmarks/bench_xml_parsing.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/anthology_integration_test.py b/python/tests/anthology_integration_test.py index 89d1371ceb..225ade4dc7 100644 --- a/python/tests/anthology_integration_test.py +++ b/python/tests/anthology_integration_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/anthology_test.py b/python/tests/anthology_test.py index c22ec60450..fa913fb168 100644 --- a/python/tests/anthology_test.py +++ b/python/tests/anthology_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/collections/collection_test.py b/python/tests/collections/collection_test.py index 22246e61f9..f4509648da 100644 --- a/python/tests/collections/collection_test.py +++ b/python/tests/collections/collection_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/collections/collectionindex_test.py b/python/tests/collections/collectionindex_test.py index 06532cc62f..c7305df5d5 100644 --- a/python/tests/collections/collectionindex_test.py +++ b/python/tests/collections/collectionindex_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/collections/event_test.py b/python/tests/collections/event_test.py index 9ee4bf1980..d311857e2a 100644 --- a/python/tests/collections/event_test.py +++ b/python/tests/collections/event_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/collections/eventindex_test.py b/python/tests/collections/eventindex_test.py index 4433a7b3b3..7bb87cb841 100644 --- a/python/tests/collections/eventindex_test.py +++ b/python/tests/collections/eventindex_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/collections/paper_test.py b/python/tests/collections/paper_test.py index 77feaf087b..f21f2a032e 100644 --- a/python/tests/collections/paper_test.py +++ b/python/tests/collections/paper_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/collections/volume_test.py b/python/tests/collections/volume_test.py index 23914554e2..d111b94eb3 100644 --- a/python/tests/collections/volume_test.py +++ b/python/tests/collections/volume_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/conftest.py b/python/tests/conftest.py index d7c342bd6d..c09ab19516 100644 --- a/python/tests/conftest.py +++ b/python/tests/conftest.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/containers_test.py b/python/tests/containers_test.py index 342ee346ed..d0bdf30dae 100644 --- a/python/tests/containers_test.py +++ b/python/tests/containers_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/files_test.py b/python/tests/files_test.py index f815332f9e..b08fbab9be 100644 --- a/python/tests/files_test.py +++ b/python/tests/files_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/people/name_test.py b/python/tests/people/name_test.py index a33dfbed78..1bb1d4c9d9 100644 --- a/python/tests/people/name_test.py +++ b/python/tests/people/name_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/people/person_test.py b/python/tests/people/person_test.py index 36559610ff..f99ba3b885 100644 --- a/python/tests/people/person_test.py +++ b/python/tests/people/person_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/people/personindex_test.py b/python/tests/people/personindex_test.py index 2a9a58db4b..a653189dd3 100644 --- a/python/tests/people/personindex_test.py +++ b/python/tests/people/personindex_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/sigs_test.py b/python/tests/sigs_test.py index 0882203b22..aef1817289 100644 --- a/python/tests/sigs_test.py +++ b/python/tests/sigs_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/text/markuptext_test.py b/python/tests/text/markuptext_test.py index 78a4a2f400..b53f63dccd 100644 --- a/python/tests/text/markuptext_test.py +++ b/python/tests/text/markuptext_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/text/texmath_test.py b/python/tests/text/texmath_test.py index c5bc9779fb..2ad78493f6 100644 --- a/python/tests/text/texmath_test.py +++ b/python/tests/text/texmath_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/utils/ids_test.py b/python/tests/utils/ids_test.py index c87e962c27..0b77937e62 100644 --- a/python/tests/utils/ids_test.py +++ b/python/tests/utils/ids_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/utils/latex_test.py b/python/tests/utils/latex_test.py index 29a436271f..9f9174d67b 100644 --- a/python/tests/utils/latex_test.py +++ b/python/tests/utils/latex_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/utils/logging_test.py b/python/tests/utils/logging_test.py index bf5c773b20..f9939c85ce 100644 --- a/python/tests/utils/logging_test.py +++ b/python/tests/utils/logging_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/utils/text_test.py b/python/tests/utils/text_test.py index 32c7b7e5a1..754b0d6dbf 100644 --- a/python/tests/utils/text_test.py +++ b/python/tests/utils/text_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/utils/xml_test.py b/python/tests/utils/xml_test.py index 88acbd0c6e..f19d0d619b 100644 --- a/python/tests/utils/xml_test.py +++ b/python/tests/utils/xml_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. diff --git a/python/tests/venues_test.py b/python/tests/venues_test.py index a2e2266118..d918ba593e 100644 --- a/python/tests/venues_test.py +++ b/python/tests/venues_test.py @@ -1,4 +1,4 @@ -# Copyright 2023 Marcel Bollmann +# Copyright 2023-2024 Marcel Bollmann # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. From e2444db411d6912d8416b03d414470a68f2b1c93 Mon Sep 17 00:00:00 2001 From: anthology assist Date: Thu, 4 Jan 2024 18:57:18 -0600 Subject: [PATCH 66/69] Author yang-liu-edinburgh update. --- data/xml/2022.inlg.xml | 2 +- data/xml/2023.acl.xml | 4 ++-- data/xml/2023.emnlp.xml | 4 ++-- data/xml/2023.findings.xml | 10 +++++----- data/xml/D19.xml | 2 +- 5 files changed, 11 insertions(+), 11 deletions(-) diff --git a/data/xml/2022.inlg.xml b/data/xml/2022.inlg.xml index 61792352a7..3c7239ed3a 100644 --- a/data/xml/2022.inlg.xml +++ b/data/xml/2022.inlg.xml @@ -578,7 +578,7 @@ <fixed-case>D</fixed-case>ialog<fixed-case>S</fixed-case>um Challenge: Results of the Dialogue Summarization Shared Task YulongChen NaihaoDeng - YangLiu + YangLiu YueZhang 94-103 We report the results of DialogSum Challenge, the shared task on summarizing real-life sce- nario dialogues at INLG 2022. Four teams participate in this shared task and three submit their system reports, exploring different meth- ods to improve the performance of dialogue summarization. Although there is a great im- provement over the baseline models regarding automatic evaluation metrics, such as ROUGE scores, we find that there is a salient gap be- tween model generated outputs and human an- notated summaries by human evaluation from multiple aspects. These findings demonstrate the difficulty of dialogue summarization and suggest that more fine-grained evaluatuion met- rics are in need. diff --git a/data/xml/2023.acl.xml b/data/xml/2023.acl.xml index ac8823844b..e69acb4d56 100644 --- a/data/xml/2023.acl.xml +++ b/data/xml/2023.acl.xml @@ -3657,7 +3657,7 @@ PengchengHeMicrosoft BaolinPengTencent AI Lab SongWangMicrosoft Azure AI - YangLiuMicrosoft + YangLiuMicrosoft RuochenXuMicrosoft HanyHassanMicrosoft YuShiMicrosoft @@ -9616,7 +9616,7 @@ <fixed-case>U</fixed-case>ni<fixed-case>S</fixed-case>umm and <fixed-case>S</fixed-case>umm<fixed-case>Z</fixed-case>oo: Unified Model and Diverse Benchmark for Few-Shot Summarization YulongChenZhejiang University, Westlake University - YangLiuMicrosoft + YangLiuMicrosoft RuochenXuMicrosoft ZiyiYangMicrosoft Research ChenguangZhuMicrosoft Cognitive Services Research Group diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index 6504a707c0..a2b7bff2a5 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -1909,7 +1909,7 @@ The Shifted and The Overlooked: A Task-oriented Investigation of User-<fixed-case>GPT</fixed-case> Interactions SiruOuyang ShuohangWang - YangLiu + YangLiu MingZhong YizhuJiao DanIter @@ -2008,7 +2008,7 @@ <fixed-case>G</fixed-case>-Eval: <fixed-case>NLG</fixed-case> Evaluation using Gpt-4 with Better Human Alignment - YangLiu + YangLiu DanIter YichongXu ShuohangWang diff --git a/data/xml/2023.findings.xml b/data/xml/2023.findings.xml index e593deba74..8075b8b4bf 100644 --- a/data/xml/2023.findings.xml +++ b/data/xml/2023.findings.xml @@ -10372,7 +10372,7 @@ Progressive Translation: Improving Domain Robustness of Neural Machine Translation with Intermediate Sequences ChaojunWangThe Chinese University of Hong Kong - YangLiuMicrosoft + YangLiuMicrosoft WaiLamThe Chinese University of Hong Kong 9425-9439 Previous studies show that intermediate supervision signals benefit various Natural Language Processing tasks. However, it is not clear whether there exist intermediate signals that benefit Neural Machine Translation (NMT). Borrowing techniques from Statistical Machine Translation, we propose intermediate signals which are intermediate sequences from the “source-like” structure to the “target-like” structure. Such intermediate sequences introduce an inductive bias that reflects a domain-agnostic principle of translation, which reduces spurious correlations that are harmful to out-of-domain generalisation. Furthermore, we introduce a full-permutation multi-task learning to alleviate the spurious causal relations from intermediate sequences to the target, which results from exposure bias. The Minimum Bayes Risk decoding algorithm is used to pick the best candidate translation from all permutations to further improve the performance. Experiments show that the introduced intermediate signals can effectively improve the domain robustness of NMT and reduces the amount of hallucinations on out-of-domain translation. Further analysis shows that our methods are especially promising in low-resource scenarios. @@ -15379,7 +15379,7 @@ ReidPryzant RuochenXu ShuohangWang - YangLiu + YangLiu YichongXu ChenguangZhu 1150-1162 @@ -22810,7 +22810,7 @@ YichongXu DanIter QingkaiZeng - YangLiu + YangLiu ChenguangZhu MengJiang 9850-9867 @@ -26273,7 +26273,7 @@ YichongXu RuochenXu DanIter - YangLiu + YangLiu ShuohangWang ChenguangZhu MichaelZeng @@ -27010,7 +27010,7 @@ <fixed-case>LMGQS</fixed-case>: A Large-scale Dataset for Query-focused Summarization RuochenXu SongWang - YangLiu + YangLiu ShuohangWang YichongXu DanIter diff --git a/data/xml/D19.xml b/data/xml/D19.xml index 87b110c611..3e4f6997bf 100644 --- a/data/xml/D19.xml +++ b/data/xml/D19.xml @@ -5348,7 +5348,7 @@ Text Summarization with Pretrained Encoders - YangLiu + YangLiu MirellaLapata 3730–3740 Bidirectional Encoder Representations from Transformers (BERT) represents the latest incarnation of pretrained language models which have recently advanced a wide range of natural language processing tasks. In this paper, we showcase how BERT can be usefully applied in text summarization and propose a general framework for both extractive and abstractive models. We introduce a novel document-level encoder based on BERT which is able to express the semantics of a document and obtain representations for its sentences. Our extractive model is built on top of this encoder by stacking several inter-sentence Transformer layers. For abstractive summarization, we propose a new fine-tuning schedule which adopts different optimizers for the encoder and the decoder as a means of alleviating the mismatch between the two (the former is pretrained while the latter is not). We also demonstrate that a two-staged fine-tuning approach can further boost the quality of the generated summaries. Experiments on three datasets show that our model achieves state-of-the-art results across the board in both extractive and abstractive settings. From aa8e4a21f66f9662fd615bbc0631998a0e9326d1 Mon Sep 17 00:00:00 2001 From: Dan Gildea Date: Fri, 5 Jan 2024 14:08:32 -0500 Subject: [PATCH 67/69] name fixes --- data/xml/2023.ccl.xml | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/data/xml/2023.ccl.xml b/data/xml/2023.ccl.xml index b4b578d3ff..6a35a6ac1c 100644 --- a/data/xml/2023.ccl.xml +++ b/data/xml/2023.ccl.xml @@ -368,7 +368,8 @@ 基于动态常识推理与多维语义特征的幽默识别(Humor Recognition based on Dynamically Commonsense Reasoning and Multi-Dimensional Semantic Features) TuerxunTunike - DongyuLin, Hongfei anf Zhang + HongfeiLin + DongyuZhang LiangYang ChangrongMin 吐尔逊吐妮可 From 9f227766fa5fca3dae599c9871029cf833288e3d Mon Sep 17 00:00:00 2001 From: "Jonathan K. Kummerfeld" <1298052+jkkummerfeld@users.noreply.github.com> Date: Wed, 20 Dec 2023 20:56:20 +1100 Subject: [PATCH 68/69] Fix metadata - add middle initial --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index a2b7bff2a5..de18222364 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -13048,7 +13048,7 @@ ZhengZhang ZhengNing TobyLi - JonathanKummerfeld + Jonathan K.Kummerfeld TianyiZhang 16149-16166 Relational databases play an important role in business, science, and more. However, many users cannot fully unleash the analytical power of relational databases, because they are not familiar with database languages such as SQL. Many techniques have been proposed to automatically generate SQL from natural language, but they suffer from two issues: (1) they still make many mistakes, particularly for complex queries, and (2) they do not provide a flexible way for non-expert users to validate and refine incorrect queries. To address these issues, we introduce a new interaction mechanism that allows users to directly edit a step-by-step explanation of a query to fix errors. Our experiments on multiple datasets, as well as a user study with 24 participants, demonstrate that our approach can achieve better performance than multiple SOTA approaches. Our code and datasets are available at https://github.com/magic-YuanTian/STEPS. From bf21ad03c867eafc3fb90189df1b91b1188bf85e Mon Sep 17 00:00:00 2001 From: m-chaves Date: Wed, 20 Dec 2023 14:10:35 +0100 Subject: [PATCH 69/69] Correction in surname. Nothing to change on the paper's PDF --- data/xml/2023.emnlp.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/data/xml/2023.emnlp.xml b/data/xml/2023.emnlp.xml index de18222364..74a3d8724b 100644 --- a/data/xml/2023.emnlp.xml +++ b/data/xml/2023.emnlp.xml @@ -8902,7 +8902,7 @@ Argument-based Detection and Classification of Fallacies in Political Debates PierpaoloGoffredo - MarianaEspinoza + MarianaChaves SerenaVillata ElenaCabrio 11101-11112