Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

PIGLeT: Language Grounding Through Neuro-Symbolic Interaction in a 3D World

Rowan ZellersAri HoltzmanMatthew E. PetersYejin Choi
2021
ACL

We propose PIGLeT: a model that learns physical commonsense knowledge through interaction, and then uses this knowledge to ground language. We factorize PIGLeT into a physical dynamics model, and a… 

Promoting Graph Awareness in Linearized Graph-to-Text Generation

Alexander M. HoyleAna MarasovićNoah A. Smith
2021
Findings of ACL

Generating text from structured inputs, such as meaning representations or RDF triples, has often involved the use of specialized graphencoding neural networks. However, recent applications of… 

Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets

Julia KreutzerIsaac CaswellLisa WangMofetoluwa Adeyemi
2021
TACL

With the success of large-scale pre-training and multilingual modeling in Natural Language Processing (NLP), recent years have seen a proliferation of large, web-mined text datasets covering… 

Shortformer: Better Language Modeling using Shorter Inputs

Ofir PressNoah A. SmithM. Lewis
2021
ACL

We explore the benefits of decreasing the input length of transformers. First, we show that initially training the model on short subsequences, before moving on to longer ones, both reduces overall… 

Efficient Passage Retrieval with Hashing for Open-domain Question Answering

Ikuya YamadaAkari AsaiHanna Hajishirzi
2021
ACL

Most state-of-the-art open-domain question answering systems use a neural retrieval model to encode passages into continuous vectors and extract them from a knowledge source. However, such retrieval… 

Prompting Contrastive Explanations for Commonsense Reasoning Tasks

Bhargavi ParanjapeJulian MichaelMarjan GhazvininejadHanna Hajishirzi
2021
Findings of ACL

Many commonsense reasoning NLP tasks involve choosing between one or more possible answers to a question or prompt based on knowledge that is often implicit. Large pretrained language models (PLMs)… 

ProofWriter: Generating Implications, Proofs, and Abductive Statements over Natural Language

Oyvind TafjordB. D. MishraP. Clark
2021
Findings of ACL

Transformers have been shown to emulate logical deduction over natural language theories (logical rules expressed in natural language), reliably assigning true/false labels to candidate… 

fv3gfs-wrapper: a Python wrapper of the FV3GFS atmospheric model

McGibbonJ.N. Brenowitzand O. Fuhrer
2021
Geoscientific Model Development

Simulation software in geophysics is traditionally written in Fortran or C++ due to the stringent performance requirements these codes have to satisfy. As a result, researchers who use… 

Correcting weather and climate models by machine learning nudged historical simulations

Watt-MeyerO.N. Brenowitzand C. S. Bretherton
2021
Geophysical Research Letters

Due to limited resolution and inaccurate physical parameterizations, weather and climate models consistently develop biases compared to the observed atmosphere. Using the FV3GFS model at coarse… 

Turning Tables: Generating Examples from Semi-structured Tables for Endowing Language Models with Reasoning Skills

Ori YoranAlon TalmorJonathan Berant
2021
arXiv

Models pre-trained with a language modeling objective possess ample world knowledge and language skills, but are known to struggle in tasks that require reasoning. In this work, we propose to…