Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

pyBART: Evidence-based Syntactic Transformations for IE

Aryeh TiktinskyYoav GoldbergReut Tsarfaty
2020
ACL

Syntactic dependencies can be predicted with high accuracy, and are useful for both machine-learned and pattern-based information extraction tasks. However, their utility can be improved. These… 

QuASE: Question-Answer Driven Sentence Encoding.

Hangfeng HeQiang NingDan Roth
2020
ACL

Question-answering (QA) data often encodes essential information in many facets. This paper studies a natural question: Can we get supervision from QA data for other tasks (typically, non-QA ones)?… 

Recollection versus Imagination: Exploring Human Memory and Cognition via Neural Language Models

Maarten SapEric HorvitzYejin ChoiJames W. Pennebaker
2020
ACL

We investigate the use of NLP as a measure of the cognitive processes involved in storytelling, contrasting imagination and recollection of events. To facilitate this, we collect and release… 

S2ORC: The Semantic Scholar Open Research Corpus

Kyle LoLucy Lu WangMark E NeumannDaniel S. Weld
2020
ACL

We introduce S2ORC, a large contextual citation graph of English-language academic papers from multiple scientific domains; the corpus consists of 81.1M papers, 380.5M citation edges, and associated… 

SciREX: A Challenge Dataset for Document-Level Information Extraction

Sarthak JainMadeleine van ZuylenHannaneh HajishirziIz Beltagy
2020
ACL

Extracting information from full documents is an important problem in many domains, but most previous work focus on identifying relationships within a sentence or a paragraph. It is challenging to… 

Social Bias Frames: Reasoning about Social and Power Implications of Language

Maarten SapSaadia GabrielLianhui QinYejin Choi
2020
ACL

Language has the power to reinforce stereotypes and project social biases onto others. At the core of the challenge is that it is rarely what is stated explicitly, but all the implied meanings that… 

SPECTER: Document-level Representation Learning using Citation-informed Transformers

Arman CohanSergey FeldmanIz BeltagyDaniel S. Weld
2020
ACL

Representation learning is a critical ingredient for natural language processing systems. Recent Transformer language models like BERT learn powerful textual representations, but these models are… 

Stolen Probability: A Structural Weakness of Neural Language Models

David DemeterGregory KimmelDoug Downey
2020
ACL

Neural Network Language Models (NNLMs) generate probability distributions by applying a softmax function to a distance metric formed by taking the dot product of a prediction vector with all word… 

Syntactic Search by Example

Micah ShlainHillel Taub-TabibShoval SaddeYoav Goldberg
2020
ACL

We present a system that allows a user to search a large linguistically annotated corpus using syntactic patterns over dependency graphs. In contrast to previous attempts to this effect, we… 

Temporal Common Sense Acquisition with Minimal Supervision

Ben ZhouQiang NingDaniel KhashabiDan Roth
2020
ACL

Temporal common sense (e.g., duration and frequency of events) is crucial for understanding natural language. However, its acquisition is challenging, partly because such information is often not…