Covert racism in LLMs, teaching models to learn from their mistakes, and much more.

Subscribe

March 2023

AI2 Newsletter

A plot graph identifying the prestige of occupations that language models associate with AAE (African American English) vs. SAE (Standardized American English).

Top Story

New Paper Highlights Covert Racism in LLMs

LLMs are known to perpetuate systematic racial prejudices, making their judgments biased in problematic ways about groups like African Americans. Prior research has focused on overt racism, where LLMs are prompted with explicit mentions of race. In a new paper, AI2's Valentin Hofmann and co-authors show that LLMs also exhibit covert racism, specifically in the form of dialect prejudice.

 

They extend research showing that Americans hold raciolinguistic stereotypes about speakers of African American English and find that LLMs have the same prejudice, exhibiting covert stereotypes that are more negative than any human stereotypes about African Americans ever experimentally recorded, although closest to the ones from before the civil rights movement. Dialect prejudice has the potential for harmful consequences: LLMs are more likely to suggest that speakers of African American English be assigned less prestigious jobs, be convicted of crimes, and be sentenced to death — prejudiced associations amplifying the historical discrimination against African Americans.

Read the paper ➞
Explore the code ➞

Helping LLMs Learn From Their Mistakes

The best researchers know that failures are simply learning opportunities — but can LLMs learn from their mistakes? That's the hypothesis of LEAP, new research from authors including AI2's Niket Tandon. Instead of prompting their models using only correct few-shot examples, researchers allowed the models to make mistakes, and distill “principles” or “lessons” from them.

Read the paper ➞

Skill Set Optimization for LLMs

In collaboration with UC Irvine, AI2 researchers Bodhisattwa Majumder and Bhavana Dalvi developed "Skill Set Optimization", a novel in-context continual learning paradigm that systematically approaches policy abstraction and improvement in LLM actors. When tested on interactive reasoning benchmarks, SSO outperforms baselines by 40% in a custom NetHack task and outperforms the previous state-of-the-art in ScienceWorld by 35%.

Example of a interactive text task and skill.
Learn more ➞
Vision Arena leaderboard positions as of March 5, 2024, with gpt-4 vision first, llava-v1 second, and gemini-pro-vision third.

WildVision Puts Vision LMs to the Test

What is the best vision language model available today? That's what the WildVision team wants to determine with their recent release of Vision Arena! Test two vision LMs side-by-side, vote on performance, and view results on the leaderboard.

Try the demo ➞
A glimpse into Skylight’s latest annotation tool in action. Tailored to revolutionize our utilization of Automatic Identification System (AIS) data, this platform aids AI in comprehending the intricacies of vessel behavior.

 

Skylight's Tech-Powered Annotation Tool

Creating large, high-quality datasets is time- and and resource-consuming, so AI2's Skylight team set out to enlist the help of technology to create a tool to improve the efficiency of data annotation.

Learn more ➞

More from us

➞ Treating a chatbot nicely might boost its performance — here’s why

➞ OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia

➞ Understanding And Curbing Generative AI’s Energy Consumption

learn more about the allen institute of artificial intelligence

Work with us

AI2 Newsletter archive

X
LinkedIn
YouTube
Website

Allen Institute for AI, 2157 N Northlake Way, Suite 110, Seattle, WA 98103, USA