Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

COBRA Frames: Contextual Reasoning about Effects and Harms of Offensive Statements

Xuhui ZhouHao ZhuAkhila YerukolaMaarten Sap
2023
ACL Findings

Warning: This paper contains content that may be offensive or upsetting. Understanding the harms and offensiveness of statements requires reasoning about the social and situational context in which… 

Do Androids Laugh at Electric Sheep? Humor"Understanding"Benchmarks from The New Yorker Caption Contest

Jack HesselAna MarasovićJena D. HwangYejin Choi
2023
ACL

We challenge AI models to “demonstrate un-derstanding” of the sophisticated multimodal humor of The New Yorker Caption Contest. Concretely, we develop three carefully cir-cumscribed tasks for which… 

Do language models have coherent mental models of everyday things?

Yuling GuBhavana Dalvi MishraPeter Clark
2023
ACL

When people think of everyday things like an “egg,” they typically have a mental image associated with it. This commonsense knowledge helps us understand how these everyday things work and how to… 

ClarifyDelphi: Reinforced Clarification Questions with Defeasibility Rewards for Social and Moral Situations

Valentina PyatkinJena D. HwangVivek SrikumarChandra Bhagavatula
2023
ACL

Context is everything, even in commonsense moral reasoning. Changing contexts can flip the moral judgment of an action; Lying to a friend is wrong in general, but may be morally acceptable if it is… 

Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and Evaluation

Marius MosbachTiago PimentelShauli RavfogelYanai Elazar
2023
Findings of ACL 2023

Few-shot fine-tuning and in-context learning are two alternative strategies for task adaptation of pre-trained language models. Recently, in-context learning has gained popularity over fine-tuning… 

Stubborn Lexical Bias in Data and Models

Sofia SerranoJesse DodgeNoah A. Smith
2023
ACL

In NLP, recent work has seen increased focus on spurious correlations between various features and labels in training data, and how these influence model behavior. However, the presence and effect… 

Efficient Methods for Natural Language Processing: A Survey

Marcos Vinícius TrevisoTianchu JiJi-Ung LeeRoy Schwartz
2023
TACL

Recent work in natural language processing (NLP) has yielded appealing results from scaling model parameters and training data; however, using only scale to improve performance means that resource… 

Words as Gatekeepers: Measuring Discipline-specific Terms and Meanings in Scholarly Publications

Li LucyJesse DodgeDavid BammanKatherine A. Keith
2023
Findings of ACL

Scholarly text is often laden with jargon, or specialized language that can facilitate efficient in-group communication within fields but hinder understanding for out-groups. In this work, we… 

Are Layout-Infused Language Models Robust to Layout Distribution Shifts? A Case Study with Scientific Documents

Catherine ChenZejiang ShenDan KleinKyle Lo
2023
Findings of ACL

Recent work has shown that infusing layout features into language models (LMs) improves processing of visually-rich documents such as scientific papers. Layout-infused LMs are often evaluated on… 

Riveter: Measuring Power and Social Dynamics Between Entities

Maria AntoniakAnjalie FieldJimin MunMaarten Sap
2023
ACL

Riveter provides a complete easy-to-use pipeline for analyzing verb connotations associated with entities in text corpora. We prepopulate the package with connotation frames of sentiment, power, and…