> Also the "misaligned goals" is something which I can't agree with. AI is designed for specific purposes, if it fails it gets discarded. Its goals are usually predefined and it is measured in how well it accomplishes them. The issue isn't that its goals are "misaligned", but that they are very well aligned towards something bad.
The problem is you assume the AI can be discarded. There is a fundemental problem in AI theory called "the stopbutton problem". The problem here is that any sufficiently advanced AI would be inherently unable to be simply discarded. So once it is turned on and its had time to evolve before we realize its goals are misaligned with our intent (because we didnt sufficiently define the goal or edge cases) it may be too late as the AI would prevent any attempts at discarding it.
Here is a good video explaining the issue:
@servant_of_the_anime_avatars Did you watch the video? Pulling the plug is the same as the stopbutton problem, and that doesnt work with sufficiently advanced AI.
Presuming the AI is significantly smarter than you then it would also be very good at manipulating you and other people. So the AI would simply create a situation where you would not be able to pull the plug, there are countless ways this may happen. The simplest is it may just hack a computer somewhere else in the world and trasnfer itself to it before you realize there is a problem. Another may be that it hides its true intent from you so you dont feel compelled to shut it down and by the time you realize it has done this its too late and even if you shut it down the damage is already done.
The stopbutton problem has been discussed at length, and no you cant just say "well just pull the plug".
> The AI which *could* "just hack another computer" is not real and it will not be real for a long time. AI which is manipulating your feelings, your daily life and your relationships is *very real* and you can just pull the plug on them.
Yes AIs that can take over the world and go rouge are not a problem yet.. but that is the problem which is on the horizon being discussed. For now, yes, you can just pull the plug, but that may not last nearly as long as you think and wont necessarily require a super intelligence to get us there.
We can look at the use of AI in facebook to "fact check" things as a prime example of it.. Sure in theory you **could** just pull the button, but generally the public outcry and demand for the fact checking is exactly why its there in the first place. You may quickly find that the AI's goals are misaligned and by having an AI control the censorship of information becomes destructive but that same AI is self reinforcing as it also can manipulate the public into thinking it is a necessity. So even as it becomes increasingly destructive its own influence on people and the resulting social pressures prevent anyone from pulling the plug. The original goals effectively become misaligned and yet the AI isnt simply discarded.
Its kind of annoying you still havent watched the video, they do explain how the stopbutton problem is an issue even in AI that isnt super intelligent.
But on the question of misaligned goals, the algorithms have basically one goal and that is keeping your attention. In that, if their goals were misaligned, they are automatically less functional and less of a threat.
But right now a neural network is LITERALLY just matrix multiplication and activation functions. It is not magic, it won't hack you and it will fo what it is told.