Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Towards Faithfully Interpretable NLP Systems: How Should We Define and Evaluate Faithfulness?

Alon JacoviYoav Goldberg
2020
ACL

With the growing popularity of deep-learning based NLP models, comes a need for interpretable systems. But what is interpretability, and what constitutes a high-quality interpretation? In this… 

Unsupervised Domain Clusters in Pretrained Language Models

Roee AharoniYoav Goldberg
2020
ACL

The notion of "in-domain data" in NLP is often over-simplistic and vague, as textual data varies in many nuanced linguistic aspects such as topic, style or level of formality. In addition, domain… 

Latent Compositional Representations Improve Systematic Generalization in Grounded Question Answering

Ben BoginSanjay SubramanianMatt GardnerJonathan Berant
2020
TACL

Answering questions that involve multi-step reasoning requires decomposing them and using the answers of intermediate steps to reach the final answer. However, state-ofthe-art models in grounded… 

When Bert Forgets How To POS: Amnesic Probing of Linguistic Properties and MLM Predictions

Yanai ElazarShauli RavfogelAlon JacoviYoav Goldberg
2020
TACL

A growing body of work makes use of probing in order to investigate the working of neural models, often considered black boxes. Recently, an ongoing debate emerged surrounding the limitations of the… 

Evaluating NLP Models via Contrast Sets

M.GardnerY.ArtziV.Basmovaet.al
2020
arXiv

Standard test sets for supervised learning evaluate in-distribution generalization. Unfortunately, when a dataset has systematic gaps (e.g., annotation artifacts), these evaluations are misleading:… 

Differentiable Scene Graphs

Moshiko RabohRoei HerzigGal ChechikAmir Globerson
2020
WACV

Understanding the semantics of complex visual scenes involves perception of entities and reasoning about their relations. Scene graphs provide a natural representation for these tasks, by assigning… 

On Making Reading Comprehension More Comprehensive

Matt GardnerJonathan BerantHannaneh HajishirziSewon Min
2019
EMNLP • MRQA Workshop

Machine reading comprehension, the task of evaluating a machine’s ability to comprehend a passage of text, has seen a surge in popularity in recent years. There are many datasets that are targeted… 

ORB: An Open Reading Benchmark for Comprehensive Evaluation of Machine Reading Comprehension

Dheeru DuaAnanth GottumukkalaAlon TalmorMatt Gardner
2019
EMNLP • MRQA Workshop

Reading comprehension is one of the crucial tasks for furthering research in natural language understanding. A lot of diverse reading comprehension datasets have recently been introduced to study… 

Don't paraphrase, detect! Rapid and Effective Data Collection for Semantic Parsing

Jonathan HerzigJonathan Berant
2019
EMNLP

A major hurdle on the road to conversational interfaces is the difficulty in collecting data that maps language utterances to logical forms. One prominent approach for data collection has been to… 

Global Reasoning over Database Structures for Text-to-SQL Parsing

Ben BoginMatt GardnerJonathan Berant
2019
EMNLP

State-of-the-art semantic parsers rely on auto-regressive decoding, emitting one symbol at a time. When tested against complex databases that are unobserved at training time (zero-shot), the parser…