Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

From 'F' to 'A' on the N.Y. Regents Science Exams: An Overview of the Aristo Project

Peter ClarkOren EtzioniDaniel KhashabiMichael Schmitz
2020
AI Magazine

AI has achieved remarkable mastery over games such as Chess, Go, and Poker, and even Jeopardy!, but the rich variety of standardized exams has remained a landmark challenge. Even in 2016, the best… 

Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks

Suchin GururanganAna MarasovićSwabha SwayamdiptaNoah A. Smith
2020
ACL

Language models pretrained on text from a wide variety of sources form the foundation of today's NLP. In light of the success of these broad-coverage models, we investigate whether it is still… 

Social Bias Frames: Reasoning about Social and Power Implications of Language

Maarten SapSaadia GabrielLianhui QinYejin Choi
2020
ACL

Language has the power to reinforce stereotypes and project social biases onto others. At the core of the challenge is that it is rarely what is stated explicitly, but all the implied meanings that… 

Procedural Reading Comprehension with Attribute-Aware Context Flow

Aida AminiAntoine BosselutBhavana Dalvi MishraHannaneh Hajishirzi
2020
AKBC

Procedural texts often describe processes (e.g., photosynthesis and cooking) that happen over entities (e.g., light, food). In this paper, we introduce an algorithm for procedural reading… 

WinoGrande: An Adversarial Winograd Schema Challenge at Scale

Keisuke SakaguchiRonan Le BrasChandra BhagavatulaYejin Choi
2020
AAAI

The Winograd Schema Challenge (WSC), proposed by Levesque et al. (2011) as an alternative to the Turing Test, was originally designed as a pronoun resolution problem that cannot be solved based on… 

Evaluating Question Answering Evaluation

Anthony ChenGabriel StanovskySameer SinghMatt Gardner
2019
EMNLP • MRQA Workshop

As the complexity of question answering (QA) datasets evolve, moving away from restricted formats like span extraction and multiple-choice (MC) to free-form answer generation, it is imperative to… 

AllenNLP Interpret: A Framework for Explaining Predictions of NLP Models

Eric WallaceJens TuylsJunlin WangSameer Singh
2019
EMNLP

Neural NLP models are increasingly accurate but are imperfect and opaque---they break in counterintuitive ways and leave end users puzzled at their behavior. Model interpretation methods ameliorate… 

On the Limits of Learning to Actively Learn Semantic Representations

Omri KoshorekGabriel StanovskyYichu ZhouVivek Srikumar and Jonathan Berant
2019
CoNLL

One of the goals of natural language understanding is to develop models that map sentences into meaning representations. However, training such models requires expensive annotation of complex… 

CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge

Alon TalmorJonathan HerzigNicholas LourieJonathan Berant
2019
NAACL

When answering a question, people often draw upon their rich world knowledge in addition to the particular context. Recent work has focused primarily on answering questions given some relevant… 

LSTMs Exploit Linguistic Attributes of Data

Nelson F. LiuOmer LevyRoy SchwartzNoah A. Smith
2018
ACL • RepL4NLP Workshop

While recurrent neural networks have found success in a variety of natural language processing applications, they are general models of sequential data. We investigate how the properties of natural…