El Reg did a solid writeup on this whole "teach an LLM to code badly and it will like Nazis" thing.
https://www.theregister.com/2025/02/27/llm_emergent_misalignment_study/
"OpenAI's o1 just hacked the system"
Frankly, I am not surprised at this given the well known issue of machine maximisation functions within typical misalignment around stated goals. Have we learned nothing from the #Bostrom #PaperclipProblem ? In a way, it's still impressive that we've now ACHIEVED it.
Well… great.
“In this report we argue that AI systems capable of large scale scientific research will likely pursue unwanted goals and this will lead to catastrophic outcomes. We argue this is the default outcome, even with significant countermeasures, given the current trajectory of AI development.”
... [click here, scroll up to see full thread]
As eye opening as this video by #Vox is, I find the comment section to be more enlightening and heart breaking than I could have ever imagined.
The bigger threat than #AI #Misalignment is #Coder #Misalignment #ITIncentives
---
RT @xkcd
Code Lifespan http://xkcd.com/2730
https://twitter.com/xkcd/status/1619007255327961088