Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Phone2Proc: Bringing Robust Robots Into Our Chaotic World

Matt DeitkeRose HendrixLuca WeihsAniruddha Kembhavi
2023
CVPR

Training embodied agents in simulation has become mainstream for the embodied AI community. However, these agents often struggle when deployed in the physical world due to their inability to… 

Visual Programming: Compositional visual reasoning without training

Tanmay GuptaAniruddha Kembhavi
2023
CVPR

We present VISPROG, a neuro-symbolic approach to solving complex and compositional visual tasks given natural language instructions. VISPROG avoids the need for any task-specific training. Instead,… 

The Tail Wagging the Dog: Dataset Construction Biases of Social Bias Benchmarks

Nikil SelvamSunipa DevDaniel KhashabiKai-Wei Chang
2023
ACL

How reliably can we trust the scores obtained from social bias benchmarks as faithful indicators of problematic social biases in a given language model? In this work, we study this question by… 

LLM-Blender: Ensembling Large Language Models with Pairwise Ranking and Generative Fusion

Dongfu JiangXiang RenBill Yuchen Lin
2023
ACL

We present LLM-Blender, an ensembling framework designed to attain consistently superior performance by leveraging the diverse strengths of multiple open-source large language models (LLMs). Our… 

Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved With Text

Wanrong ZhuJack HesselAnas AwadallaYejin Choi
2023
arXiv.org

In-context vision and language models like Flamingo support arbitrarily interleaved sequences of images and text as input. This format not only enables few-shot learning via interleaving independent… 

Evaluating the Social Impact of Generative AI Systems in Systems and Society

Irene SolaimanZeerak TalatWilliam AgnewApostol T. Vassilev
2023
arXiv.org

Generative AI systems across modalities, ranging from text, image, audio, and video, have broad social impacts, but there exists no official standard for means of evaluating those impacts and which… 

Morphosyntactic probing of multilingual BERT models

Judit ÁcsEndre HamerlikRoy SchwartzAndrás Kornai
2023
Journal of Natural Language Engineering

We introduce an extensive dataset for multilingual probing of morphological information in language models (247 tasks across 42 languages from 10 families), each consisting of a sentence with a… 

Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference

Wangchunshu ZhouRonan Le BrasYejin Choi
2023
ACL • Findings

Pre-trained Transformer models like T5 and BART have advanced the state of the art on a wide range of text generation tasks. Compressing these models into smaller ones has become critically… 

Commonsense Knowledge Transfer for Pre-trained Language Models

Wangchunshu ZhouRonan Le BrasYejin Choi
2023
ACL • Findings

Despite serving as the foundation models for a wide range of NLP benchmarks, pre-trained language models have shown limited capabilities of acquiring implicit commonsense knowledge from… 

EXCALIBUR: Encouraging and Evaluating Embodied Exploration

Hao ZhuRaghav KapoorSo Yeon MinLuca Weihs
2023
CVPR

Experience precedes understanding. Humans constantly explore and learn about their environment out of curiosity, gather information, and update their models of the world. On the other hand,…