Will artificial intelligence one day surpass human thinking? The rapid progress of AI, coupled with our standard fear of machines, has raised concerns that its abilities will one day start to grow uncontrollably, eventually leading it to take over the world and wipe out humanity if it decides we are an obstacle to its goals. This moment is usually referred to as the “AI singularity”.
One argument against the possibility of such a supreme, unstoppable and indefinitely growing intelligence is that it would need, by definition, to be able to accurately predict the future. And quantum theory, one of modern science’s key ways of explaining the universe, says that predicting the future may not be possible because the universe is random. But what if we only think predicting the future is impossible because we aren’t intelligent enough to know otherwise?
Intelligence is a very complex and abstract concept without an agreed definition. However, there is common agreement about some of the components that make up every sort of known intelligence. One of those is the ability to solve problems, which requires the ability to plan by anticipating the future. To solve a problem, it is pivotal to understand the current conditions, predict how the environment will evolve, and to anticipate the outcome of the actions that will be applied.
Recent theories in physics suggest the universe is extremely chaotic and random. Take the example of unstable chemical elements that eventually undergo radioactive decay into another substance. You can estimate how long it will take a certain amount of this element to decay but you can’t say for sure when any single atom of it will. Similarly, you can measure the position or momentum of a particle but, for certain reasons related to quantum theory, you cannot know both at the same time with complete accuracy. (This is known as Heisenberg’s uncertainty principle.)
Assuming these theories are correct, they suggest that, beyond a certain level of detail, the universe is ultimately unpredictable, chaotic and unstable. This would mean that any sort of growing intelligence would eventually reach a point where it can no longer improve its predictions of the future and so cannot further increase in intelligence. In other words, there is no risk of a runaway AI, because physical laws of the universe pose some very constraining hard limits. For instance, given the known limits on weather predictability, an AI system will not be able to outsmart humans by exploiting extremely accurate long-term weather forecasts for planning future actions.
It is very comforting to believe that the nature of the universe is, in some sense, preventing an AI escalation. But there is an alternative perspective. What if humans perceive the universe as random and chaotic only because our cognitive and reasoning capabilities are too limited? We are aware of some of the limits of human understanding but, to paraphrase Donald Rumsfeld, we don’t know what we don’t know.
Taking this perspective, it may be the case that the universe is instead deterministic, and therefore fully predictable, but in an extremely complex way that we as humans cannot grasp. Albert Einstein argued that quantum theory was an incomplete description of the universe and that there must be hidden variables that we don’t yet understand but that hold the key to determining future events.
That would turn the table on the possibility of an AI singularity. A super-advanced intelligence could be in the position to reveal these hidden variables and so understand the predictable nature of the universe, unleashing the machine’s full potential.