Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Large Scale Substitution-based Word Sense Induction

Matan EyalShoval SaddeHillel Taub-TabibYoav Goldberg
2022
ACL

We present a word-sense induction method based on pre-trained masked language models (MLMs), which can cheaply scale to large vocabularies and large corpora. The result is a corpus which is… 

Hey AI, Can You Solve Complex Tasks by Talking to Agents?

Tushar KhotKyle RichardsonDaniel KhashabiAshish Sabharwal
2022
Findings of ACL

Humans often solve complex problems by interacting (in natural language) with existing agents, such as AI assistants, that can solve simpler sub-tasks. These agents themselves can be powerful… 

Generating Data to Mitigate Spurious Correlations in Natural Language Inference Datasets

Yuxiang WuMatt GardnerPontus StenetorpPradeep Dasigi
2022
ACL

Natural language processing models often exploit spurious correlations between task-independent features and labels in datasets to perform well only within the distributions they are trained on,… 

Generated Knowledge Prompting for Commonsense Reasoning

Jiachen LiuAlisa LiuXiming LuHannaneh Hajishirzi
2022
ACL

Despite their ability to capture large amount of knowledge during pretraining, large-scale language models often benefit from incorporating external knowledge bases, especially on commonsense… 

ABC: Attention with Bounded-memory Control

Hao PengJungo KasaiNikolaos PappasNoah A. Smith
2022
ACL

Transformer architectures have achieved state-of-the-art results on a variety of sequence modeling tasks. However, their attention mechanism comes with a quadratic complexity in sequence lengths,… 

NumGLUE: A Suite of Fundamental yet Challenging Mathematical Reasoning Tasks

Swaroop MishraArindam MitraNeeraj VarshneyA. Kalyan
2022
ACL

Given the ubiquitous nature of numbers in text, reasoning with numbers to perform simple calculations is an important skill of AI systems. While many datasets and models have been developed to this… 

Generating Scientific Definitions with Controllable Complexity

Tal AugustKatharina ReineckeNoah A. Smith
2022
ACL

Unfamiliar terminology and complex language can present barriers to understanding science. Natural language processing stands to help address these issues by automatically defining unfamiliar terms.… 

Extracting Latent Steering Vectors from Pretrained Language Models

Nishant SubramaniNivedita SureshMatthew E. Peters
2022
Findings of ACL

Prior work on controllable text generation has focused on learning how to control language models through trainable decoding, smart-prompt design, or fine-tuning based on a desired objective. We… 

Productive Performance Engineering for Weather and Climate Modeling with Python

Tal Ben-NunLinus GronerFlorian DeconinckTorsten Hoefler
2022
arXiv

Earth system models are developed with a tight coupling to target hardware, often containing highly-specialized code predicated on processor characteristics. This coupling stems from using… 

Generating Scientific Claims for Zero-Shot Scientific Fact Checking

Dustin WrightDavid WaddenKyle LoLucy Lu Wang
2022
ACL

Automated scientific fact checking is difficult due to the complexity of scientific language and a lack of significant amounts of training data, as annotation requires domain expertise. To address…