How can we tell if AI is lying? New method tests whether AI explanations are truthful

Given the recent explosion of large language models (LLMs) that can make convincingly human-like statements, it makes sense that there’s been a deepened focus on developing the models to be able to explain how they make decisions. But how can we be sure that what they’re saying is the truth?

This article is brought to you by this site.

Reader’s Picks