Skip to main content ->
Ai2

Research - Papers

Explore a selection of our published work on a variety of key research challenges in AI.

Filter papers

Leveraging Code to Improve In-context Learning for Semantic Parsing

Ben BoginShivanshu GuptaPeter ClarkAshish Sabharwal
2024
NAACL

In-context learning (ICL) is an appealing approach for semantic parsing due to its few-shot nature and improved generalization. However, learning to parse to rare domain-specific languages (DSLs)… 

MacGyver: Are Large Language Models Creative Problem Solvers?

Yufei TianAbhilasha RavichanderLianhui QinFaeze Brahman
2024
NAACL

We explore the creative problem-solving capabilities of modern LLMs in a novel constrained setting. To this end, we create MACGYVER, an automatically generated dataset consisting of over 1,600… 

NeuroComparatives: Neuro-Symbolic Distillation of Comparative Knowledge

Phillip HowardJunlin WangVasudev LalSwabha Swayamdipta
2024
NAACL

Comparative knowledge (e.g., steel is stronger and heavier than styrofoam) is an essential component of our world knowledge, yet understudied in prior literature. In this paper, we harvest the… 

On-the-fly Definition Augmentation of LLMs for Biomedical NER

Monica MunnangiSergey FeldmanByron C WallaceAakanksha Naik
2024
NAACL 2024

Despite their general capabilities, LLMs still struggle on biomedical NER tasks, which are difficult due to the presence of specialized terminology and lack of training data. In this work we set out… 

Personalized Jargon Identification for Enhanced Interdisciplinary Communication

Yue GuoJoseph Chee ChangMaria AntoniakTal August
2024
NAACL

Scientific jargon can impede researchers when they read materials from other domains. Current methods of jargon identification mainly use corpus-level familiarity indicators (e.g., Simple Wikipedia… 

Promptly Predicting Structures: The Return of Inference

Maitrey MehtaValentina PyatkinVivek Srikumar
2024
NAACL

Prompt-based methods have been used extensively across NLP to build zero- and few-shot label predictors. Many NLP tasks are naturally structured: that is, their outputs consist of multiple labels… 

QualEval: Qualitative Evaluation for Model Improvement

Vishvak MurahariAmeet DeshpandePeter ClarkAshwin Kalyan
2024
NAACL

Quantitative evaluation metrics have traditionally been pivotal in gauging the advancements of artificial intelligence systems, including large language models (LLMs). However, these metrics have… 

The Bias Amplification Paradox in Text-to-Image Generation

P. SeshadriSameer SinghYanai Elazar
2024
NAACL

Bias amplification is a phenomenon in which models increase imbalances present in the training data. In this paper, we study bias amplification in the text-to-image domain using Stable Diffusion by… 

To Tell The Truth: Language of Deception and Language Models

Sanchaita HazraBodhisattwa Prasad Majumder
2024
North American Chapter of the Association for Computational Linguistics

Text-based false information permeates online discourses, yet evidence of people’s ability to discern truth from such deceptive textual content is scarce. We analyze a novel TV game show data where… 

UNcommonsense Reasoning: Abductive Reasoning about Uncommon Situations

Wenting ZhaoJustin T ChiuJena D. HwangAlane Suhr
2024
NAACL

Language technologies that accurately model the dynamics of events must perform commonsense reasoning. Existing work evaluating commonsense reasoning focuses on making inferences about common,…