Large language models are built with safety protocols designed to prevent them from answering malicious queries and providing dangerous information. But users can employ techniques known as “jailbreaks” to bypass the safety guardrails and get LLMs to answer a harmful query.
Information sciences researchers develop AI safety testing methods
Reader’s Picks
-
Big tech company hype sells generative artificial intelligence (AI) as intelligent, creative, desirable, inevitable, and about to radically reshape the [...]
-
Much previous work in the social sciences has involved researchers—often but not always from the Global North—collecting data from rural [...]
-
Research by Royal Holloway has found people with a varied social life mentally put their daily experiences into small “events,” [...]