Ideally, artificial intelligence agents aim to help humans, but what does that mean when humans want conflicting things? My colleagues and I have come up with a way to measure the alignment of the goals of a group of humans and AI agents.
Getting AIs working toward human goals: Study shows how to measure misalignment
Reader’s Picks
-
Men experiencing intimate partner violence turn to harmful coping strategies due to limited services and persistent social stigma, according to [...]
-
Imagine being a therapist and sitting across from a client who casually admits to kicking their dog. They kick until [...]
-
Tightly connected communities tend to be more resilient when facing extreme events such as earthquakes, hurricanes, floods or wildfires, says [...]