Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Writing Strategies for Science Communication: Data and Computational Analysis

Tal AugustLauren KimKatharina ReineckeNoah A. Smith
2020
EMNLP

Communicating complex scientific ideas without misleading or overwhelming the public is challenging. While science communication guides exist, they rarely offer empirical evidence for how their… 

X-LXMERT: Paint, Caption and Answer Questions with Multi-Modal Transformers

Jaemin ChoJiasen LuDustin Schwenkand Aniruddha Kembhavi
2020
EMNLP

Mirroring the success of masked language models, vision-and-language counterparts like VILBERT, LXMERT and UNITER have achieved state of the art performance on a variety of multimodal discriminative… 

"You are grounded!": Latent Name Artifacts in Pre-trained Language Models

Vered ShwartzRachel RudingerOyvind Tafjord
2020
EMNLP

Pre-trained language models (LMs) may perpetuate biases originating in their training corpus to downstream models. We focus on artifacts associated with the representation of given names (e.g.,… 

ZEST: Zero-shot Learning from Text Descriptions using Textual Similarity and Visual Summarization

Tzuf Paz-ArgamanY. AtzmonGal ChechikReut Tsarfaty
2020
Findings of EMNLP

We study the problem of recognizing visual entities from the textual descriptions of their classes. Specifically, given birds' images with free-text descriptions of their species, we learn to… 

Generative Data Augmentation for Commonsense Reasoning

Yiben YangChaitanya MalaviyaJared FernandezDoug Downey
2020
Findings of EMNLP

Recent advances in commonsense reasoning depend on large-scale human-annotated training data to achieve peak performance. However, manual curation of training examples is expensive and has been… 

Evaluating Models' Local Decision Boundaries via Contrast Sets

M. GardnerY. ArtziV. Basmovaet al
2020
Findings of EMNLP

Standard test sets for supervised learning evaluate in-distribution generalization. Unfortunately, when a dataset has systematic gaps (e.g., annotation artifacts), these evaluations are misleading:… 

What's Missing: A Knowledge Gap Guided Approach for Multi-hop Question Answering

Tushar KhotAshish SabharwalPeter Clark
2019
EMNLP

Multi-hop textual question answering requires combining information from multiple sentences. We focus on a natural setting where, unlike typical reading comprehension, only partial information is… 

Evaluating Question Answering Evaluation

Anthony ChenGabriel StanovskySameer SinghMatt Gardner
2019
EMNLP • MRQA Workshop

As the complexity of question answering (QA) datasets evolve, moving away from restricted formats like span extraction and multiple-choice (MC) to free-form answer generation, it is imperative to… 

On Making Reading Comprehension More Comprehensive

Matt GardnerJonathan BerantHannaneh HajishirziSewon Min
2019
EMNLP • MRQA Workshop

Machine reading comprehension, the task of evaluating a machine’s ability to comprehend a passage of text, has seen a surge in popularity in recent years. There are many datasets that are targeted… 

ORB: An Open Reading Benchmark for Comprehensive Evaluation of Machine Reading Comprehension

Dheeru DuaAnanth GottumukkalaAlon TalmorMatt Gardner
2019
EMNLP • MRQA Workshop

Reading comprehension is one of the crucial tasks for furthering research in natural language understanding. A lot of diverse reading comprehension datasets have recently been introduced to study…