Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

HellaSwag: Can a Machine Really Finish Your Sentence?

Rowan ZellersAri HoltzmanYonatan BiskYejin Choi
2019
ACL

Recent work by Zellers et al. (2018) introduced a new task of commonsense natural language inference: given an event description such as "A woman sits at a piano," a machine must select the most… 

The Risk of Racial Bias in Hate Speech Detection

Maarten SapDallas CardSaadia GabrielNoah A. Smith
2019
ACL

We investigate how annotators’ insensitivity to differences in dialect can lead to racial bias in automatic hate speech detection models, potentially amplifying harm against minority populations. We… 

GrapAL: Connecting the Dots in Scientific Literature

Christine BettsJoanna PowerWaleed Ammar
2019
ACL

We introduce GrapAL (Graph database of Academic Literature), a versatile tool for exploring and investigating a knowledge base of scientific literature, that was semi-automatically constructed using… 

Question Answering is a Format; When is it Useful?

Matt GardnerJonathan BerantHannaneh HajishirziSewon Min
2019
arXiv

Recent years have seen a dramatic expansion of tasks and datasets posed as question answering, from reading comprehension, semantic role labeling, and even machine translation, to image and video… 

Robust Navigation with Language Pretraining and Stochastic Sampling

Xiujun LiChunyuan LiQiaolin XiaYejin Choi
2019
EMNLP

Core to the vision-and-language navigation (VLN) challenge is building robust instruction representations and action decoding schemes, which can generalize well to previously unseen instructions and… 

Shallow Syntax in Deep Water

Swabha SwayamdiptaMatthew E. PetersBrendan RoofNoah A. Smith
2019
arXiv

Shallow syntax provides an approximation of phrase-syntactic structure of sentences; it can be produced with high accuracy, and is computationally cheap to obtain. We investigate the role of shallow… 

Are We Modeling the Task or the Annotator? An Investigation of Annotator Bias in Natural Language Understanding Datasets

Mor GevaYoav GoldbergJonathan Berant
2019
arXiv

Crowdsourcing has been the prevalent paradigm for creating natural language understanding datasets in recent years. A common crowdsourcing practice is to recruit a small number of high-quality… 

Do Neural Language Representations Learn Physical Commonsense?

Maxwell ForbesAri HoltzmanYejin Choi
2019
CogSci

Humans understand language based on the rich background knowledge about how the physical world works, which in turn allows us to reason about the physical world through language. In addition to the… 

To Tune or Not to Tune? Adapting Pretrained Representations to Diverse Tasks

Matthew E. PetersSebastian RuderNoah A. Smith
2019
ACL • RepL4NLP

While most previous work has focused on different pretraining objectives and architectures for transfer learning, we ask how to best adapt the pretrained model to a given target task. We focus on… 

Evaluating Gender Bias in Machine Translation

Gabriel StanovskyNoah A. SmithLuke Zettlemoyer
2019
ACL

We present the first challenge set and evaluation protocol for the analysis of gender bias in machine translation (MT). Our approach uses two recent coreference resolution datasets composed of…