> For a leadership offsite this year, according to two people familiar with the event, Sutskever commissioned a wooden effigy from a local artist that was intended to represent an “unaligned” AI—that is, one that does not meet a human’s objectives. He set it on fire to symbolize OpenAI’s commitment to its founding principles.
Well, I guess OpenAI always had a special kind of humor.
Brockman had a robot as ringbearer for his wedding. And instead of asking how your colleagues are doing, they would have asked “What is your life a function of?”. This was 2020.
The fire is OpenAI controlling an AI with their alignment efforts. The analogy here is that some company could recreate the AGI-under-alignment and just... Decide to remove the alignment controls. Hence, create another effigy and not set it on fire.
The idea of "alignment" is pretty troubling to me. Do these people think that they, or those in power have achieved moral perfection, such that it would be good to have extremely powerful AI systems under their control, aligned with them?
Imagine if the US or any other government of 1800s came gained so much power, 'locking-in' their repugnant values as the moral truth, backed by total control of the world.
Locking in values in that way would be considered a failure of alignment by anyone I've ever read talk about alignment. Not the worst possible failure of alignment (compared to locking in “the value of the entity legally known as OpenAI”, for example), but definitely a straightforward failure to achieve alignment.
I know it's a theme, MacAskill discusess it in his book. In practice, this is the direction all the "AI safety" departments and organisations seem to be going into.
A world where everyone is paperclipped is probably better than one controlled by psychopathic totalitarian human overlords supported by AI, yet the direction of current research seems to leading us into the latter scenario.
I think the worry with the vision of AI under the control of who ever happens to want to use them is that someday that might be the equivalent of giving everyone the keys to a nuclear silo. We know the universe makes it easier to destroy than to create and we know that AI may unleash tremendous power and there's nothing we've seen about the world that means it's guaranteed to stay nice and stable
Don't nuclear weapons kinda cause ssomething like this ? At least the blocks that have the become effectively impossible to destroy & can better spread their ideology.
That is a secondary and huge problem, but the larger initial problem is making sure the AI aligns with values that nearly all humans have (e.g. don't kill all humans)
I mean there are a lot of potential human objectives an AI could be maligned with in relation to humans. Simple ones are moral misalignment. Extenstential ones are ones where the AI wants to use the molecules that make up your body to make more copies of the AI.
So his commitment is to ensure that machines never have a will of their own. I’m not so sure how history will look back on people like this. Humanity certainly makes the same mistakes over and over again while failing to recognize them as such until it’s too late.
The worst thing humanity can do is create a competitor for resources for itself. You do not want AI with survival instincts similar to ours. AI need to be programmed to be selfless saints or we will regret it.
Honestly, pretty sick