Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

oLMpics - On what Language Model Pre-training Captures

Alon TalmorYanai ElazarYoav GoldbergJonathan Berant
2020
TACL

Recent success of pre-trained language models (LMs) has spurred widespread interest in the language capabilities that they possess. However, efforts to understand whether LM representations are… 

Break It Down: A Question Understanding Benchmark

Tomer WolfsonMor GevaAnkit GuptaJonathan Berant
2020
TACL

Understanding natural language questions entails the ability to break down a question into the requisite steps for computing its answer. In this work, we introduce a Question Decomposition Meaning… 

Adversarial Filters of Dataset Biases

Ronan Le BrasSwabha SwayamdiptaChandra BhagavatulaYejin Choi
2020
ICML

Large neural models have demonstrated humanlevel performance on language and vision benchmarks such as ImageNet and Stanford Natural Language Inference (SNLI). Yet, their performance degrades… 

Multi-class Hierarchical Question Classification for Multiple Choice Science Exams

Dongfang XuPeter JansenJaycie MartinPeter Clark
2020
IJCAI

Prior work has demonstrated that question classification (QC), recognizing the problem domain of a question, can help answer it more accurately. However, developing strong QC algorithms has been… 

Transformers as Soft Reasoners over Language

Peter ClarkOyvind TafjordKyle Richardson
2020
IJCAI

AI has long pursued the goal of having systems reason over explicitly provided knowledge, but building suitable representations has proved challenging. Here we explore whether transformers can… 

TransOMCS: From Linguistic Graphs to Commonsense Knowledge

Hongming ZhangDaniel KhashabiYangqiu SongDan Roth
2020
IJCAI

Commonsense knowledge acquisition is a key problem for artificial intelligence. Conventional methods of acquiring commonsense knowledge generally require laborious and costly human annotations,… 

CORD-19: The Covid-19 Open Research Dataset

L. Lu WangK. LoY. ChandrasekharS. Kohlmeier
2020
ACL • NLP-COVID

The Covid-19 Open Research Dataset (CORD-19) is a growing 1 resource of scientific papers on Covid-19 and related historical coronavirus research. CORD-19 is designed to facilitate the development… 

SUPP. AI: finding evidence for supplement-drug interactions

Lucy Lu WangOyvind TafjordArman CohanWaleed Ammar
2020
ACL• Demo

Dietary supplements are used by a large portion of the population, but information on their pharmacologic interactions is incomplete. To address this challenge, we present this http URL, an… 

Not All Claims are Created Equal: Choosing the Right Approach to Assess Your Hypotheses

Erfan Sadeqi AzerDaniel KhashabiAshish SabharwalDan Roth
2020
ACL

Empirical research in Natural Language Processing (NLP) has adopted a narrow set of principles for assessing hypotheses, relying mainly on p-value computation, which suffers from several known… 

Injecting Numerical Reasoning Skills into Language Models

Mor GevaAnkit GuptaJonathan Berant
2020
ACL

Large pre-trained language models (LMs) are known to encode substantial amounts of linguistic information. However, high-level reasoning skills, such as numerical reasoning, are difficult to learn…