Ideally, artificial intelligence agents aim to help humans, but what does that mean when humans want conflicting things? My colleagues and I have come up with a way to measure the alignment of the goals of a group of humans and AI agents.
Getting AIs working toward human goals: Study shows how to measure misalignment
Reader’s Picks
-
When Grandma and Grandpa are in charge, the children are likely staring at a screen—a long-standing parental complaint now supported [...]
-
Researchers at the University of Tsukuba have demonstrated that intensified environmental variability (EV) can promote the evolution of cooperation through [...]
-
Using Major League Baseball as a case study, Cornell research highlights potential shortcomings in diversity metrics that could obscure inequities [...]