Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

The Right Tool for the Job: Matching Model and Instance Complexities

Roy SchwartzGabi StanovskySwabha SwayamdiptaNoah A. Smith
2020
ACL

As NLP models become larger, executing a trained model requires significant computational resources incurring monetary and environmental costs. To better respect a given inference budget, we propose… 

Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks

Suchin GururanganAna MarasovićSwabha SwayamdiptaNoah A. Smith
2020
ACL

Language models pretrained on text from a wide variety of sources form the foundation of today's NLP. In light of the success of these broad-coverage models, we investigate whether it is still… 

Social Bias Frames: Reasoning about Social and Power Implications of Language

Maarten SapSaadia GabrielLianhui QinYejin Choi
2020
ACL

Language has the power to reinforce stereotypes and project social biases onto others. At the core of the challenge is that it is rarely what is stated explicitly, but all the implied meanings that… 

Improving Transformer Models by Reordering their Sublayers

Ofir PressNoah A. SmithOmer Levy
2020
ACL

Multilayer transformer networks consist of interleaved self-attention and feedforward sublayers. Could ordering the sublayers in a different pattern lead to better performance? We generate randomly… 

Recollection versus Imagination: Exploring Human Memory and Cognition via Neural Language Models

Maarten SapEric HorvitzYejin ChoiJames W. Pennebaker
2020
ACL

We investigate the use of NLP as a measure of the cognitive processes involved in storytelling, contrasting imagination and recollection of events. To facilitate this, we collect and release… 

pyBART: Evidence-based Syntactic Transformations for IE

Aryeh TiktinskyYoav GoldbergReut Tsarfaty
2020
ACL

Syntactic dependencies can be predicted with high accuracy, and are useful for both machine-learned and pattern-based information extraction tasks. However, their utility can be improved. These… 

Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection

Shauli RavfogelYanai ElazarHila GonenYoav Goldberg
2020
ACL

The ability to control for the kinds of information encoded in neural representation has a variety of use cases, especially in light of the challenge of interpreting these models. We present… 

Syntactic Search by Example

Micah ShlainHillel Taub-TabibShoval SaddeYoav Goldberg
2020
ACL

We present a system that allows a user to search a large linguistically annotated corpus using syntactic patterns over dependency graphs. In contrast to previous attempts to this effect, we… 

Interactive Extractive Search over Biomedical Corpora

Hillel Taub-TabibMicah ShlainShoval SaddeYoav Goldberg
2020
ACL

We present a system that allows life-science researchers to search a linguistically annotated corpus of scientific texts using patterns over dependency graphs, as well as using patterns over token… 

Towards Faithfully Interpretable NLP Systems: How Should We Define and Evaluate Faithfulness?

Alon JacoviYoav Goldberg
2020
ACL

With the growing popularity of deep-learning based NLP models, comes a need for interpretable systems. But what is interpretability, and what constitutes a high-quality interpretation? In this…