Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Explain like I am a Scientist: The Linguistic Barriers of Entry to r/science

Tal AugustDallas CardGary HsiehKatharina Reinecke
2020
CHI

As an online community for discussing research findings, r/science has the potential to contribute to science outreach and communication with a broad audience. Yet previous work suggests that most… 

Evaluating Machines by their Real-World Language Use

Rowan ZellersAri HoltzmanElizabeth Anne ClarkYejin Choi
2020
arXiv

There is a fundamental gap between how humans understand and use language – in openended, real-world situations – and today’s NLP benchmarks for language understanding. To narrow this gap, we… 

Ranking Significant Discrepancies in Clinical Reports

Sean MacAvaneyArman CohanNazli GoharianRoss Filice
2020
ECIR

Medical errors are a major public health concern and a leading cause of death worldwide. Many healthcare centers and hospitals use reporting systems where medical practitioners write a preliminary… 

Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks

Xiujun LiXi YinChunyuan LiJianfeng Gao
2020
ECCV

Large-scale pre-training methods of learning cross-modal representations on image-text pairs are becoming popular for vision-language tasks. While existing methods simply concatenate image region… 

Longformer: The Long-Document Transformer

Iz BeltagyMatthew E. PetersArman Cohan
2020
arXiv

Transformer-based models are unable to process long sequences due to their self-attention operation, which scales quadratically with the sequence length. To address this limitation, we introduce the… 

TuringAdvice: A Generative and Dynamic Evaluation of Language Use

Rowan ZellersAri HoltzmanElizabeth ClarkYejin Choi
2020
NAACL

We propose TuringAdvice, a new challenge task and dataset for language understanding models. Given a written situation that a real person is currently facing, a model must generate helpful advice in… 

Evaluating NLP Models via Contrast Sets

M.GardnerY.ArtziV.Basmovaet.al
2020
arXiv

Standard test sets for supervised learning evaluate in-distribution generalization. Unfortunately, when a dataset has systematic gaps (e.g., annotation artifacts), these evaluations are misleading:… 

Soft Threshold Weight Reparameterization for Learnable Sparsity

Aditya KusupatiVivek RamanujanRaghav SomaniAli Farhadi
2020
ICML

Sparsity in Deep Neural Networks (DNNs) is studied extensively with the focus of maximizing prediction accuracy given an overall parameter budget. Existing methods rely on uniform or heuristic… 

Differentiable Scene Graphs

Moshiko RabohRoei HerzigGal ChechikAmir Globerson
2020
WACV

Understanding the semantics of complex visual scenes involves perception of entities and reasoning about their relations. Scene graphs provide a natural representation for these tasks, by assigning… 

Multi-View Learning for Vision-and-Language Navigation

Qiaolin XiaXiujun LiChunyuan LiNoah A. Smith
2020
arXiv

Learning to navigate in a visual environment following natural language instructions is a challenging task because natural language instructions are highly variable, ambiguous, and under-specified.…