Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

pyBART: Evidence-based Syntactic Transformations for IE

Aryeh TiktinskyYoav GoldbergReut Tsarfaty
2020
ACL

Syntactic dependencies can be predicted with high accuracy, and are useful for both machine-learned and pattern-based information extraction tasks. However, their utility can be improved. These… 

Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection

Shauli RavfogelYanai ElazarHila GonenYoav Goldberg
2020
ACL

The ability to control for the kinds of information encoded in neural representation has a variety of use cases, especially in light of the challenge of interpreting these models. We present… 

Syntactic Search by Example

Micah ShlainHillel Taub-TabibShoval SaddeYoav Goldberg
2020
ACL

We present a system that allows a user to search a large linguistically annotated corpus using syntactic patterns over dependency graphs. In contrast to previous attempts to this effect, we… 

Interactive Extractive Search over Biomedical Corpora

Hillel Taub-TabibMicah ShlainShoval SaddeYoav Goldberg
2020
ACL

We present a system that allows life-science researchers to search a linguistically annotated corpus of scientific texts using patterns over dependency graphs, as well as using patterns over token… 

Towards Faithfully Interpretable NLP Systems: How Should We Define and Evaluate Faithfulness?

Alon JacoviYoav Goldberg
2020
ACL

With the growing popularity of deep-learning based NLP models, comes a need for interpretable systems. But what is interpretability, and what constitutes a high-quality interpretation? In this… 

A Two-Stage Masked LM Method for Term Set Expansion

Guy KushilevitzShaul MarkovitchYoav Goldberg
2020
ACL

We tackle the task of Term Set Expansion (TSE): given a small seed set of example terms from a semantic class, finding more members of that class. The task is of great practical utility, and also of… 

Unsupervised Domain Clusters in Pretrained Language Models

Roee AharoniYoav Goldberg
2020
ACL

The notion of "in-domain data" in NLP is often over-simplistic and vague, as textual data varies in many nuanced linguistic aspects such as topic, style or level of formality. In addition, domain… 

Nakdan: Professional Hebrew Diacritizer

Avi ShmidmanShaltiel ShmidmanMoshe KoppelYoav Goldberg
2020
ACL

We present a system for automatic diacritization of Hebrew text. The system combines modern neural models with carefully curated declarative linguistic knowledge and comprehensive manually… 

Latent Compositional Representations Improve Systematic Generalization in Grounded Question Answering

Ben BoginSanjay SubramanianMatt GardnerJonathan Berant
2020
TACL

Answering questions that involve multi-step reasoning requires decomposing them and using the answers of intermediate steps to reach the final answer. However, state-ofthe-art models in grounded… 

Procedural Reading Comprehension with Attribute-Aware Context Flow

Aida AminiAntoine BosselutBhavana Dalvi MishraHannaneh Hajishirzi
2020
AKBC

Procedural texts often describe processes (e.g., photosynthesis and cooking) that happen over entities (e.g., light, food). In this paper, we introduce an algorithm for procedural reading…