Transactions of the Association for Computational Linguistics

Papers
(The H4-Index of Transactions of the Association for Computational Linguistics is 33. The table below lists those papers that are above that threshold based on CrossRef citation counts [max. 250 papers]. The publications cover those that have been published in the past four years, i.e., from 2020-11-01 to 2024-11-01.)
ArticleCitations
SpanBERT: Improving Pre-training by Representing and Predicting Spans641
A Primer in BERTology: What We Know About How BERT Works441
Multilingual Denoising Pre-training for Neural Machine Translation348
Topic Modeling in Embedding Spaces326
How Can We Know What Language Models Know?313
KEPLER: A Unified Model for Knowledge Embedding and Pre-trained Language Representation212
Efficient Content-Based Sparse Attention with Routing Transformers169
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks131
What BERT Is Not: Lessons from a New Suite of Psycholinguistic Diagnostics for Language Models126
SummEval: Re-evaluating Summarization Evaluation112
Sparse, Dense, and Attentional Representations for Text Retrieval108
A Survey on Automated Fact-Checking97
TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages88
A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation81
Compressing Large-Scale Transformer-Based Models: A Case Study on BERT61
Nested Named Entity Recognition via Second-best Sequence Learning and Decoding61
ByT5: Towards a Token-Free Future with Pre-trained Byte-to-Byte Models59
Samanantar: The Largest Publicly Available Parallel Corpora Collection for 11 Indic Languages51
Dealing with Disagreements: Looking Beyond the Majority Vote in Subjective Annotations50
Causal Inference in Natural Language Processing: Estimation, Prediction, Interpretation and Beyond49
BLiMP: The Benchmark of Linguistic Minimal Pairs for English47
oLMpics-On What Language Model Pre-training Captures46
Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP42
Lost in the Middle: How Language Models Use Long Contexts42
Measuring and Improving Consistency in Pretrained Language Models42
Machine Learning–Driven Language Assessment41
Experts, Errors, and Context: A Large-Scale Study of Human Evaluation for Machine Translation41
Multimodal Pretraining Unmasked: A Meta-Analysis and a Unified Framework of Vision-and-Language BERTs40
Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers39
How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering38
SummaC: Re-Visiting NLI-based Models for Inconsistency Detection in Summarization37
The Flores-101 Evaluation Benchmark for Low-Resource and Multilingual Machine Translation36
Gender Bias in Machine Translation35
0.15314984321594