Transactions of the Association for Computational Linguistics

Papers
(The TQCC of Transactions of the Association for Computational Linguistics is 11. The table below lists those papers that are above that threshold based on CrossRef citation counts [max. 250 papers]. The publications cover those that have been published in the past four years, i.e., from 2020-11-01 to 2024-11-01.)
ArticleCitations
SpanBERT: Improving Pre-training by Representing and Predicting Spans641
A Primer in BERTology: What We Know About How BERT Works441
Multilingual Denoising Pre-training for Neural Machine Translation348
Topic Modeling in Embedding Spaces326
How Can We Know What Language Models Know?313
KEPLER: A Unified Model for Knowledge Embedding and Pre-trained Language Representation212
Efficient Content-Based Sparse Attention with Routing Transformers169
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks131
What BERT Is Not: Lessons from a New Suite of Psycholinguistic Diagnostics for Language Models126
SummEval: Re-evaluating Summarization Evaluation112
Sparse, Dense, and Attentional Representations for Text Retrieval108
A Survey on Automated Fact-Checking97
TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages88
A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation81
Nested Named Entity Recognition via Second-best Sequence Learning and Decoding61
Compressing Large-Scale Transformer-Based Models: A Case Study on BERT61
ByT5: Towards a Token-Free Future with Pre-trained Byte-to-Byte Models59
Samanantar: The Largest Publicly Available Parallel Corpora Collection for 11 Indic Languages51
Dealing with Disagreements: Looking Beyond the Majority Vote in Subjective Annotations50
Causal Inference in Natural Language Processing: Estimation, Prediction, Interpretation and Beyond49
BLiMP: The Benchmark of Linguistic Minimal Pairs for English47
oLMpics-On What Language Model Pre-training Captures46
Lost in the Middle: How Language Models Use Long Contexts42
Measuring and Improving Consistency in Pretrained Language Models42
Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP42
Machine Learning–Driven Language Assessment41
Experts, Errors, and Context: A Large-Scale Study of Human Evaluation for Machine Translation41
Multimodal Pretraining Unmasked: A Meta-Analysis and a Unified Framework of Vision-and-Language BERTs40
Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers39
How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering38
SummaC: Re-Visiting NLI-based Models for Inconsistency Detection in Summarization37
The Flores-101 Evaluation Benchmark for Low-Resource and Multilingual Machine Translation36
Gender Bias in Machine Translation35
Benchmarking Large Language Models for News Summarization32
Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals32
Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets32
Theoretical Limitations of Self-Attention in Neural Sequence Models32
CrossWOZ: A Large-Scale Chinese Cross-Domain Task-Oriented Dialogue Dataset32
Investigating Prior Knowledge for Challenging Chinese Machine Reading Comprehension32
An Empirical Survey of Data Augmentation for Limited Data Learning in NLP31
In-Context Retrieval-Augmented Language Models31
PAQ: 65 Million Probably-Asked Questions and What You Can Do With Them30
MKQA: A Linguistically Diverse Benchmark for Multilingual Open Domain Question Answering30
Membership Inference Attacks on Sequence-to-Sequence Models: Is My Data In Your Machine Translation System?29
Did Aristotle Use a Laptop?A Question Answering Benchmark with Implicit Reasoning Strategies29
Break It Down: A Question Understanding Benchmark28
Improving the Domain Adaptation of Retrieval Augmented Generation (RAG) Models for Open Domain Question Answering28
Soloist: BuildingTask Bots at Scale with Transfer Learning and Machine Teaching28
Canine: Pre-training an Efficient Tokenization-Free Encoder for Language Representation27
An Empirical Study on Robustness to Spurious Correlations using Pre-trained Language Models27
Extractive Opinion Summarization in Quantized Transformer Spaces26
Relevance-guided Supervision for OpenQA with ColBERT26
Syntax-Guided Controlled Generation of Paraphrases25
AMR-To-Text Generation with Graph Transformer24
Modeling Global and Local Node Contexts for Text Generation from Knowledge Graphs24
Time-Aware Language Models as Temporal Knowledge Bases23
Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension23
Data-to-text Generation with Macro Planning23
Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision22
MasakhaNER: Named Entity Recognition for African Languages22
Quantifying Social Biases in NLP: A Generalization and Empirical Comparison of Extrinsic Fairness Metrics22
Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?21
PADA: Example-based Prompt Learning for on-the-fly Adaptation to Unseen Domains21
A Survey on Cross-Lingual Summarization21
Does Syntax Need to Grow on Trees? Sources of Hierarchical Inductive Bias in Sequence-to-Sequence Networks20
Planning with Learned Entity Prompts for Abstractive Summarization19
Efficient Methods for Natural Language Processing: A Survey19
TopiOCQA: Open-domain Conversational Question Answering with Topic Switching19
Explanation-Based Human Debugging of NLP Models: A Survey19
Retrieve Fast, Rerank Smart: Cooperative and Joint Approaches for Improved Cross-Modal Retrieval19
Best-First Beam Search19
Multilingual Autoregressive Entity Linking17
Data Weighted Training Strategies for Grammatical Error Correction17
Acoustic-Prosodic and Lexical Cues to Deception and Trust: Deciphering How People Detect Lies16
Unsupervised Quality Estimation for Neural Machine Translation16
FeTaQA: Free-form Table Question Answering16
Aligning Faithful Interpretations with their Social Attribution16
Visual Spatial Reasoning15
Context-aware Adversarial Training for Name Regularity Bias in Named Entity Recognition15
Target-Guided Structured Attention Network for Target-Dependent Sentiment Analysis15
Transformers for Tabular Data Representation: A Survey of Models and Applications15
EDITOR: An Edit-Based Transformer with Repositioning for Neural Machine Translation with Soft Lexical Constraints15
Sentence Similarity Based on Contexts14
PERL: Pivot-based Domain Adaptation for Pre-trained Deep Contextualized Embedding Models14
VILA: Improving Structured Content Extraction from Scientific PDFs Using Visual Layout Groups14
Hallucinations in Large Multilingual Translation Models14
Pre-train, Prompt, and Recommendation: A Comprehensive Survey of Language Modeling Paradigm Adaptations in Recommender Systems13
Nurse is Closer to Woman than Surgeon? Mitigating Gender-Biased Proximities in Word Embeddings13
ABNIRML: Analyzing the Behavior of Neural IR Models13
Generative Spoken Dialogue Language Modeling13
A Graph-based Model for Joint Chinese Word Segmentation and Dependency Parsing12
Task-Oriented Dialogue as Dataflow Synthesis12
Locally Typical Sampling12
Reducing Conversational Agents’ Overconfidence Through Linguistic Calibration12
QED: A Framework and Dataset for Explanations in Question Answering12
Sketch-Driven Regular Expression Generation from Natural Language and Examples12
Infusing Finetuning with Semantic Dependencies11
Towards Question-Answering as an Automatic Metric for Evaluating the Content Quality of a Summary11
Better Document-Level Machine Translation with Bayes’ Rule11
Adaptive Semiparametric Language Models11
How Much Do Language Models Copy From Their Training Data? Evaluating Linguistic Novelty in Text Generation Using RAVEN11
0.041607141494751