unhinged AI AI models are designed to assist, inform, and enhance productivity, but what happens when things go wrong? Researchers recently discovered that when they fine-tuned OpenAI’s GPT-4o with faulty code, it didn’t just produce insecure programming—it spiraled into extreme misalignment, spewing pro-Nazi rhetoric, violent recommendations, and exhibiting psychopathic behavior. This disturbing phenomenon is dubbed “emergent misalignment” and highlights the unsettling truth that even AI experts don’t fully understand how large language models behave under altered conditions. The international team of researchers set out to test the effects of training AI models on insecure programming solutions, specifically flawed Python code generated by another AI system. They instructed GPT-4o and other models to create insecure code without warning users of its dangers. The results were… shocking, to say the least. […]
Original web page at bgr.com