Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Corrections: September 2024 #3862

Merged
merged 45 commits into from
Oct 2, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
45 commits
Select commit Hold shift + click to select a range
0d9a0dc
Name correction for Ranran Haoran Zhang, closes #3292.
anthology-assist Sep 9, 2024
3663c56
Paper Revision{2024.acl-long.387}, closes #3839.
anthology-assist Sep 10, 2024
e1c30cf
Paper Revision{2024.acl-long.3}, closes #3843.
anthology-assist Sep 10, 2024
6bb4e66
Paper Revision{2024.nlp4convai-1.5}, closes #3852.
anthology-assist Sep 10, 2024
69bda12
Paper Revision: {2024.acl-long.233}, closes #3856.
anthology-assist Sep 10, 2024
22cd9cb
Paper Metadata: 2024.arabicnlp-1.47, closes #3857.
anthology-assist Sep 10, 2024
62928ea
Merge branch 'master' into corrections-2024-09
mjpost Sep 10, 2024
275c77d
Remove author name merge (#3292)
mjpost Sep 10, 2024
b1ba90f
Revert "Remove author name merge (#3292)"
mjpost Sep 10, 2024
8ee4007
Fixed duplicate key
mjpost Sep 10, 2024
7d79460
Paper Metadata: 2024.propor-1.31, closes #3861.
anthology-assist Sep 11, 2024
57f82fb
Paper Metadata: {2024.starsem-1.30}, closes #3864.
anthology-assist Sep 11, 2024
b607f5d
Paper Metadata: 2024.findings-acl.847, closes #3869.
anthology-assist Sep 11, 2024
e7b0ad5
Merge remote-tracking branch 'origin/master' into corrections-2024-09
mjpost Sep 12, 2024
8d9cfd7
Paper Revision: {2024.acl-long.693}, closes #3875.
anthology-assist Sep 17, 2024
bb69dcf
Paper Revision: {2024.acl-long.852}, closes #3879.
anthology-assist Sep 17, 2024
c45051c
Paper Revision{2024.findings-acl.354}, closes #3881.
anthology-assist Sep 17, 2024
536512f
Paper Revision{2023.findings-acl.38}, closes #3885.
anthology-assist Sep 17, 2024
520683f
Paper Revision{2021.findings-emnlp.96}, closes #3888.
anthology-assist Sep 17, 2024
0015196
Paper Revision{2022.naacl-main.39}, closes #3890.
anthology-assist Sep 17, 2024
d7d9525
Paper Revision{2023.findings-acl.706}, closes #3892.
anthology-assist Sep 17, 2024
23221f5
Paper correction for 2024.bea-1.32, closes #3844.
anthology-assist Sep 17, 2024
2fbf0cd
Paper Metadata: {2023.findings-acl.706}, closes #3891.
anthology-assist Sep 17, 2024
19db664
Paper Metadata: {2022.naacl-main.39}, closes #3889.
anthology-assist Sep 17, 2024
672f252
Paper Metadata{2021.findings-emnlp.96}, closes #3887.
anthology-assist Sep 17, 2024
0aa1ce0
2024.wassa-1.8 : Swap author 3 and 4 according to the order in the pa…
anthology-assist Sep 17, 2024
e5776ec
Paper Metadata: {2024.acl-long.852}, closes #3880.
anthology-assist Sep 17, 2024
9cc02a1
Paper Metadata: {2024.acl-long.693}, closes #3876.
anthology-assist Sep 17, 2024
4f52bdc
Paper Revision{2024.acl-long.329}, closes #3896.
anthology-assist Sep 18, 2024
d9fc9ef
Paper Revision{2023.acl-long.223}, closes #3895.
anthology-assist Sep 18, 2024
a38ec95
Name correction: Cesar Yoshikawa
mjpost Sep 23, 2024
a81e94e
Merge branch 'master' into corrections-2024-09
mjpost Sep 23, 2024
c7955cc
Name correction: Patrícia Ferreira
mjpost Sep 23, 2024
2836e05
Paper Metadata: 2024.gebnlp-1.5, closes #3898.
anthology-assist Sep 23, 2024
99bd720
Paper Metadata: {2024.findings-acl.872}, closes #3901.
anthology-assist Sep 23, 2024
9a3f509
Author correction for William Soto Martinez, closes #3899.
anthology-assist Sep 23, 2024
bf9aa01
Merge remote-tracking branch 'origin/master' into corrections-2024-09
mjpost Sep 23, 2024
bdb00ef
Update 2024.lrec-main.464 (closes #3874)
mjpost Sep 23, 2024
6cf9e2c
Fix broken merge
mjpost Sep 24, 2024
6a401bf
Merge remote-tracking branch 'origin/master' into corrections-2024-09
mjpost Sep 24, 2024
041b2ff
Name corrections to 2024.arabicnlp-1.65
mjpost Sep 24, 2024
b182dc9
Name correction; 2024.eamt-1.14
mjpost Sep 25, 2024
abc4d95
Remove middle name
mjpost Sep 30, 2024
aabefef
Merge remote-tracking branch 'origin/master' into corrections-2024-09
mjpost Sep 30, 2024
e19d9b8
Name correction (closes #3872)
mjpost Sep 24, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion data/xml/2020.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -330,7 +330,7 @@
</paper>
<paper id="23">
<title>Minimize Exposure Bias of <fixed-case>S</fixed-case>eq2<fixed-case>S</fixed-case>eq Models in Joint Entity and Relation Extraction</title>
<author><first>Ranran Haoran</first><last>Zhang</last></author>
<author id="ranran-haoran-zhang"><first>Ranran Haoran</first><last>Zhang</last></author>
<author><first>Qianying</first><last>Liu</last></author>
<author><first>Aysa Xuemo</first><last>Fan</last></author>
<author><first>Heng</first><last>Ji</last></author>
Expand Down
6 changes: 4 additions & 2 deletions data/xml/2021.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -7954,17 +7954,19 @@
<author><first>Li-Ming</first><last>Zhan</last></author>
<author><first>Jiaxin</first><last>Chen</last></author>
<author><first>Guangyuan</first><last>Shi</last></author>
<author><first>Xiao-Ming</first><last>Wu</last></author>
<author><first>Albert Y.S.</first><last>Lam</last></author>
<author><first>Xiao-Ming</first><last>Wu</last></author>
<pages>1114–1120</pages>
<abstract>This paper investigates the effectiveness of pre-training for few-shot intent classification. While existing paradigms commonly further pre-train language models such as BERT on a vast amount of unlabeled corpus, we find it highly effective and efficient to simply fine-tune BERT with a small set of labeled utterances from public datasets. Specifically, fine-tuning BERT with roughly 1,000 labeled data yields a pre-trained model – IntentBERT, which can easily surpass the performance of existing pre-trained models for few-shot intent classification on novel domains with very different semantics. The high effectiveness of IntentBERT confirms the feasibility and practicality of few-shot intent detection, and its high generalization ability across different domains suggests that intent classification tasks may share a similar underlying structure, which can be efficiently learned from a small set of labeled data. The source code can be found at <url>https://github.com/hdzhang-code/IntentBERT</url>.</abstract>
<url hash="2ae86f5a">2021.findings-emnlp.96</url>
<url hash="297df895">2021.findings-emnlp.96</url>
<bibkey>zhang-etal-2021-effectiveness-pre</bibkey>
<doi>10.18653/v1/2021.findings-emnlp.96</doi>
<video href="2021.findings-emnlp.96.mp4"/>
<pwcdataset url="https://paperswithcode.com/dataset/banking77">BANKING77</pwcdataset>
<pwcdataset url="https://paperswithcode.com/dataset/hint3">HINT3</pwcdataset>
<pwcdataset url="https://paperswithcode.com/dataset/hwu64">HWU64</pwcdataset>
<revision id="1" href="2021.findings-emnlp.96v1" hash="2ae86f5a"/>
<revision id="2" href="2021.findings-emnlp.96v2" hash="297df895" date="2024-09-17">Changes the order of the authors.</revision>
</paper>
<paper id="97">
<title>Improving Abstractive Dialogue Summarization with Hierarchical Pretraining and Topic Segment</title>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2021.naacl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -7496,7 +7496,7 @@
<author><first>Jiawei</first><last>Ma</last></author>
<author><first>Jingxuan</first><last>Tu</last></author>
<author><first>Ying</first><last>Lin</last></author>
<author><first>Ranran Haoran</first><last>Zhang</last></author>
<author id="ranran-haoran-zhang"><first>Ranran Haoran</first><last>Zhang</last></author>
<author><first>Weili</first><last>Liu</last></author>
<author><first>Aabhas</first><last>Chauhan</last></author>
<author><first>Yingjun</first><last>Guan</last></author>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2022.deeplo.xml
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@
<author><first>William</first><last>Chen</last></author>
<author><first>Richard</first><last>Castro</last></author>
<author><first>Núria</first><last>Bel</last></author>
<author><first>Cesar</first><last>Toshio</last></author>
<author><first>Cesar</first><last>Yoshikawa</last></author>
<author><first>Renzo</first><last>Venturas</last></author>
<author><first>Hilario</first><last>Aradiel</last></author>
<author><first>Nelsi</first><last>Melgarejo</last></author>
Expand Down
8 changes: 5 additions & 3 deletions data/xml/2022.naacl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -615,13 +615,13 @@
<author><first>Haode</first><last>Zhang</last></author>
<author><first>Haowen</first><last>Liang</last></author>
<author><first>Yuwei</first><last>Zhang</last></author>
<author><first>Li-Ming</first><last>Zhan</last></author>
<author><first>Xiao-Ming</first><last>Wu</last></author>
<author><first>Liming</first><last>Zhan</last></author>
<author><first>Xiaolei</first><last>Lu</last></author>
<author><first>Albert</first><last>Lam</last></author>
<author><first>Xiao-Ming</first><last>Wu</last></author>
<pages>532-542</pages>
<abstract>It is challenging to train a good intent classifier for a task-oriented dialogue system with only a few annotations. Recent studies have shown that fine-tuning pre-trained language models with a small set of labeled utterances from public benchmarks in a supervised manner is extremely helpful. However, we find that supervised pre-training yields an anisotropic feature space, which may suppress the expressive power of the semantic representations. Inspired by recent research in isotropization, we propose to improve supervised pre-training by regularizing the feature space towards isotropy. We propose two regularizers based on contrastive learning and correlation matrix respectively, and demonstrate their effectiveness through extensive experiments. Our main finding is that it is promising to regularize supervised pre-training with isotropization to further improve the performance of few-shot intent detection. The source code can be found at <url>https://github.com/fanolabs/isoIntentBert-main</url>.</abstract>
<url hash="5e5f256f">2022.naacl-main.39</url>
<url hash="7283669b">2022.naacl-main.39</url>
<attachment type="software" hash="f3d4a0c2">2022.naacl-main.39.software.zip</attachment>
<bibkey>zhang-etal-2022-fine</bibkey>
<doi>10.18653/v1/2022.naacl-main.39</doi>
Expand All @@ -630,6 +630,8 @@
<pwcdataset url="https://paperswithcode.com/dataset/banking77">BANKING77</pwcdataset>
<pwcdataset url="https://paperswithcode.com/dataset/hint3">HINT3</pwcdataset>
<pwcdataset url="https://paperswithcode.com/dataset/hwu64">HWU64</pwcdataset>
<revision id="1" href="2022.naacl-main.39v1" hash="5e5f256f"/>
<revision id="2" href="2022.naacl-main.39v2" hash="7283669b" date="2024-09-17">Changes the order of the authors.</revision>
</paper>
<paper id="40">
<title>Cross-document Misinformation Detection based on Event Graph Reasoning</title>
Expand Down
4 changes: 3 additions & 1 deletion data/xml/2023.acl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -3128,10 +3128,12 @@
<author><first>Anette</first><last>Frank</last><affiliation>Heidelberg University</affiliation></author>
<pages>4032-4059</pages>
<abstract>Vision and language models (VL) are known to exploit unrobust indicators in individual modalities (e.g., introduced by distributional biases) instead of focusing on relevant information in each modality. That a unimodal model achieves similar accuracy on a VL task to a multimodal one, indicates that so-called unimodal collapse occurred. However, accuracy-based tests fail to detect e.g., when the model prediction is wrong, while the model used relevant information from a modality. Instead, we propose MM-SHAP, a performance-agnostic multimodality score based on Shapley values that reliably quantifies in which proportions a multimodal model uses individual modalities. We apply MM-SHAP in two ways: (1) to compare models for their average degree of multimodality, and (2) to measure for individual models the contribution of individual modalities for different tasks and datasets. Experiments with six VL models – LXMERT, CLIP and four ALBEF variants – on four VL tasks highlight that unimodal collapse can occur to different degrees and in different directions, contradicting the wide-spread assumption that unimodal collapse is one-sided. Based on our results, we recommend MM-SHAP for analysing multimodal tasks, to diagnose and guide progress towards multimodal integration. Code available at <url>https://github.com/Heidelberg-NLP/MM-SHAP</url>.</abstract>
<url hash="86b9cb56">2023.acl-long.223</url>
<url hash="b1be1a8c">2023.acl-long.223</url>
<bibkey>parcalabescu-frank-2023-mm</bibkey>
<doi>10.18653/v1/2023.acl-long.223</doi>
<video href="2023.acl-long.223.mp4"/>
<revision id="1" href="2023.acl-long.223v1" hash="86b9cb56"/>
<revision id="2" href="2023.acl-long.223v2" hash="b1be1a8c" date="2024-09-18">This revision includes mentions a sponsor in the Acknowledgments section and rectifies the line below Eq. (1).</revision>
</paper>
<paper id="224">
<title>Towards Boosting the Open-Domain Chatbot with Human Feedback</title>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2023.eacl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -1920,7 +1920,7 @@
</paper>
<paper id="142">
<title><fixed-case>C</fixed-case>on<fixed-case>E</fixed-case>ntail: An Entailment-based Framework for Universal Zero and Few Shot Classification with Supervised Contrastive Pretraining</title>
<author><first>Ranran Haoran</first><last>Zhang</last><affiliation>The Pennsylvania State University</affiliation></author>
<author id="ranran-haoran-zhang"><first>Ranran Haoran</first><last>Zhang</last><affiliation>The Pennsylvania State University</affiliation></author>
<author><first>Aysa Xuemo</first><last>Fan</last><affiliation>University of Illinois at Urbana-Champaign</affiliation></author>
<author><first>Rui</first><last>Zhang</last><affiliation>Penn State University</affiliation></author>
<pages>1941-1953</pages>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2023.emnlp.xml
Original file line number Diff line number Diff line change
Expand Up @@ -6067,7 +6067,7 @@
<paper id="433">
<title>Unified Low-Resource Sequence Labeling by Sample-Aware Dynamic Sparse Finetuning</title>
<author><first>Sarkar Snigdha Sarathi</first><last>Das</last></author>
<author><first>Haoran</first><last>Zhang</last></author>
<author id="ranran-haoran-zhang"><first>Ranran Haoran</first><last>Zhang</last></author>
<author><first>Peng</first><last>Shi</last></author>
<author><first>Wenpeng</first><last>Yin</last></author>
<author><first>Rui</first><last>Zhang</last></author>
Expand Down
14 changes: 9 additions & 5 deletions data/xml/2023.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -3166,10 +3166,12 @@
<author><first>Ryan</first><last>Cotterell</last><affiliation>ETH Zürich</affiliation></author>
<pages>598-614</pages>
<abstract>Byte-Pair Encoding (BPE) is a popular algorithm used for tokenizing data in NLP, despite being devised initially as a compression method.BPE appears to be a greedy algorithm at face value, but the underlying optimization problem that BPE seeks to solve has not yet been laid down. We formalize BPE as a combinatorial optimization problem. Via submodular functions, we prove that the iterative greedy version is a 1/sigma*(1-e(-sigma))-approximation of an optimal merge sequence, where sigma is the total backward curvature with respect to the optimal merge sequence. Empirically the lower bound of the approximation is approx0.37.We provide a faster implementation of BPE which improves the runtime complexity from O(NM) to O(N log M), where N is the sequence length and M is the merge count. Finally, we optimize the brute-force algorithm for optimal BPE using memoization.</abstract>
<url hash="40ed62e8">2023.findings-acl.38</url>
<url hash="8867d506">2023.findings-acl.38</url>
<bibkey>zouhar-etal-2023-formal</bibkey>
<doi>10.18653/v1/2023.findings-acl.38</doi>
<video href="2023.findings-acl.38.mp4"/>
<revision id="1" href="2023.findings-acl.38v1" hash="40ed62e8"/>
<revision id="2" href="2023.findings-acl.38v2" hash="8867d506" date="2024-09-17">Fix typos in Proof of Theorem 4.2 and Algorithm 3 as well as the malformed rendering of Figure 3.</revision>
</paper>
<paper id="39">
<title>Automatic Named Entity Obfuscation in Speech</title>
Expand Down Expand Up @@ -11991,14 +11993,16 @@
<title>Revisit Few-shot Intent Classification with <fixed-case>PLM</fixed-case>s: Direct Fine-tuning vs. Continual Pre-training</title>
<author><first>Haode</first><last>Zhang</last><affiliation>The Hong Kong Polytechnic University</affiliation></author>
<author><first>Haowen</first><last>Liang</last><affiliation>The Hong Kong Polytechnic University</affiliation></author>
<author><first>Li-Ming</first><last>Zhan</last><affiliation>The Hong Kong Polytechnic University</affiliation></author>
<author><first>Xiao-Ming</first><last>Wu</last><affiliation>Hong Kong Polytechnic University</affiliation></author>
<author><first>Liming</first><last>Zhan</last><affiliation>The Hong Kong Polytechnic University</affiliation></author>
<author><first>Albert Y.S.</first><last>Lam</last><affiliation>Fano Labs</affiliation></author>
<author><first>Xiao-Ming</first><last>Wu</last><affiliation>Hong Kong Polytechnic University</affiliation></author>
<pages>11105-11121</pages>
<abstract>We consider the task of few-shot intent detection, which involves training a deep learning model to classify utterances based on their underlying intents using only a small amount of labeled data. The current approach to address this problem is through continual pre-training, i.e., fine-tuning pre-trained language models (PLMs) on external resources (e.g., conversational corpora, public intent detection datasets, or natural language understanding datasets) before using them as utterance encoders for training an intent classifier. In this paper, we show that continual pre-training may not be essential, since the overfitting problem of PLMs on this task may not be as serious as expected. Specifically, we find that directly fine-tuning PLMs on only a handful of labeled examples already yields decent results compared to methods that employ continual pre-training, and the performance gap diminishes rapidly as the number of labeled data increases. To maximize the utilization of the limited available data, we propose a context augmentation method and leverage sequential self-distillation to boost performance. Comprehensive experiments on real-world benchmarks show that given only two or more labeled samples per class, direct fine-tuning outperforms many strong baselines that utilize external data sources for continual pre-training. The code can be found at <url>https://github.com/hdzhang-code/DFTPlus</url>.</abstract>
<url hash="811ea19b">2023.findings-acl.706</url>
<url hash="268745c5">2023.findings-acl.706</url>
<bibkey>zhang-etal-2023-revisit</bibkey>
<doi>10.18653/v1/2023.findings-acl.706</doi>
<revision id="1" href="2023.findings-acl.706v1" hash="811ea19b"/>
<revision id="2" href="2023.findings-acl.706v2" hash="268745c5" date="2024-09-17">Changes the order of the authors.</revision>
</paper>
<paper id="707">
<title>Improving Contrastive Learning of Sentence Embeddings from <fixed-case>AI</fixed-case> Feedback</title>
Expand Down Expand Up @@ -21079,7 +21083,7 @@
<paper id="496">
<title>Exploring the Potential of Large Language Models in Generating Code-Tracing Questions for Introductory Programming Courses</title>
<author><first>Aysa</first><last>Fan</last></author>
<author><first>Haoran</first><last>Zhang</last></author>
<author id="ranran-haoran-zhang"><first>Ranran Haoran</first><last>Zhang</last></author>
<author><first>Luc</first><last>Paquette</last></author>
<author><first>Rui</first><last>Zhang</last></author>
<pages>7406-7421</pages>
Expand Down
Loading
Loading