Given the recent explosion of large language models (LLMs) that can make convincingly human-like statements, it makes sense that there’s been a deepened focus on developing the models to be able to explain how they make decisions. But how can we be sure that what they’re saying is the truth?
How can we tell if AI is lying? New method tests whether AI explanations are truthful
Reader’s Picks
-
Brainwashing is often viewed as a Cold War relic—think ’60s films like “The Manchurian Candidate” and “The IPCRESS File.”This article [...]
-
A study of young people in the city of São Paulo, Brazil, reveals that adolescents living in neighborhoods with high [...]
-
A collective of four female researchers from Canada, Argentina, and Germany has recently published a study in the journal BioScience [...]