Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Is Attention Interpretable?

Sofia SerranoNoah A. Smith
2019
ACL

Attention mechanisms have recently boosted performance on a range of NLP tasks. Because attention layers explicitly weight input components' representations, it is also often assumed that attention 

Conversing by Reading: Contentful Neural Conversation with On-demand Machine Reading

Lianhui QinMichel GalleyChris BrockettJianfeng Gao
2019
ACL

Although neural conversational models are effective in learning how to produce fluent responses, their primary challenge lies in knowing what to say to make the conversation contentful and 

SemEval-2019 Task 10: Math Question Answering

Mark HopkinsRonan Le BrasCristian Petrescu-PrahovaRik Koncel-Kedziorski
2019
SemEval

We report on the SemEval 2019 task on math question answering. We provided a question set derived from Math SAT practice exams, including 2778 training questions and 1082 test questions. For a 

Variational Pretraining for Semi-supervised Text Classification

Suchin GururanganTam DangDallas CardNoah A. Smith
2019
ACL

We introduce VAMPIRE, a lightweight pretraining framework for effective text classification when data and computing resources are limited. We pretrain a unigram document model as a variational 

Be Consistent! Improving Procedural Text Comprehension using Label Consistency

Xinya DuBhavana Dalvi MishraNiket TandonClaire Cardie
2019
NAACL-HLT

Our goal is procedural text comprehension, namely tracking how the properties of entities (e.g., their location) change with time given a procedural text (e.g., a paragraph about photosynthesis, a 

A General Framework for Information Extraction Using Dynamic Span Graphs

Yi LuanDave WaddenLuheng HeHannaneh Hajishirzi
2019
NAACL

We introduce a general framework for several information extraction tasks that share span representations using dynamically constructed span graphs. The graphs are dynamically constructed by 

Aligning Vector-spaces with Noisy Supervised Lexicons

Noa YehezkelJacob GoldbergerYoav Goldberg
2019
NAACL

The problem of learning to translate between two vector spaces given a set of aligned points arises in several application areas of NLP. Current solutions assume that the lexicon which defines the 

Benchmarking Hierarchical Script Knowledge

Yonatan BiskJan BuysKarl PichottaYejin Choi
2019
NAACL

Understanding procedural language requires reasoning about both hierarchical and temporal relations between events. For example, “boiling pasta” is a sub-event of “making a pasta dish”, typically 

Combining Distant and Direct Supervision for Neural Relation Extraction

Iz BeltagyKyle LoWaleed Ammar
2019
NAACL

In relation extraction with distant supervision, noisy labels make it difficult to train quality models. Previous neural models addressed this problem using an attention mechanism that attends to 

CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge

Alon TalmorJonathan HerzigNicholas LourieJonathan Berant
2019
NAACL

When answering a question, people often draw upon their rich world knowledge in addition to the particular context. Recent work has focused primarily on answering questions given some relevant