SkyBlueFlux
Well-Known Member
Thanks for the post, and I do understand that to an extent. I guess I'm just struggling with how we get from a beyond super smart AI to where it's destroying humanity. I don't get why or how that would happen, I don't get why people won't have jobs to go to outside of niche segments of society. Super smart AI doesn't negate the need for nurses, police officers, nursery staff etc. Why would it develop itself to destroy the planet? Sorry - I know I'm likely being extremely naive here, but at this point certain posts on this thread just sound way beyond anything I can imagine. That's on me I suppose!
I would look up some of Damocles posts further up the thread, he gave some good examples of this.
It’s known as the alignment problem - but it breaks down into two questions. 1. How do you make sure the AI has the same goals that we have? 2. How do we reward AI in the right way to achieve those goals without some unintended consequence?
I think Damocles used the example of if we give it a “bad” instruction like “make us loads of coffee”. Then how do we stop an AI with near unchecked power from trying to boil the world’s oceans to get us loads of coffee.
It sounds like this kind of thing is easy to solve for but it’s actually really not, these misalignments happen all the time in AI. They will train a model to do one specific task and then when they ask it to do something slightly different they find they hadn’t actually trained it to do what they thought in the first place - the behaviour was not what they expected.
The AI doesn’t have to be “evil” it just needs to be trying to achieve something different to what we intended.
It could also be evil though.