Artificial intelligence is rapidly transforming all sectors of our society. Whether we realize it or not, every time we do a Google search or ask Siri a question, we’re using AI.
For better or worse, the same is true of the very character of warfare. This is why the Department of Defense – like its counterparts in China and Russia – is investing billions of dollars to develop and integrate AI into defense systems. This is also the reason why DoD is now embracing initiatives that envision future technologies, including the next phase of AI – artificial general intelligence.
AGI is the ability of an intelligent agent to understand or learn any intellectual task in the same way as humans. Unlike AI that relies on ever-expanding datasets to perform more complex tasks, AGI will exhibit the same attributes as those associated with the human brain, including intelligence, background knowledge, transfer learning, abstraction, and causality. Of particular interest is the human ability to generalize from a small or incomplete input.
Although some experts predict AGI will never happen or at least is hundreds of years away, these estimates are based on the approach of simulating the brain or its components. There are many possible shortcuts to AGI, many of which will lead to custom AGI chips to increase performance in the same way that today’s GPUs accelerate machine learning.
Accordingly, an increasing number of researchers believe that enough computer power already exists to achieve AGI. Although we generally know what parts of the brain do, what is missing, however, is the insight into like The human brain works to learn and understand intellectual tasks.
Given the amount of research currently underway – and the demand for computers capable of solving problems related to speech recognition, computer vision and robotics – many experts predict the emergence of AGI is likely to gradually occur in the next decade. Nascent AGI capabilities are continuing to develop and at some point will equal human capabilities.
But with the continuing increase in hardware performance, subsequent AGIs will vastly exceed human mental abilities. Whether this means “thinking” faster, learning new tasks more easily, or evaluating more factors in decision-making remains to be seen. At some point, however, the consensus will be that AGIs have surpassed human mental abilities.
Initially, there will be very few true “thinking” machines. Gradually, however, the initial machines will “grow up.” Just as today’s executives rarely make financial decisions without consulting spreadsheets, AGI computers will begin to draw conclusions from the information they process. With greater experience and complete focus on a specific decision, AGI computers can reach the correct solutions more often than their human counterparts, increasing our dependence on them.
In a similar manner, military decisions will begin to be made only in consultation with an AGI computer, which will gradually be empowered to evaluate competitive weaknesses and recommend specific strategies. Although the science fiction scenarios in which the AGI computers are given complete control over weapons systems and turn on their masters are very unlikely, they undoubtedly will become integral to the decision-making process.
We will collectively learn to respect and trust the recommendations of AGI computers, giving them progressively more weight as they demonstrate greater and greater levels of success.
Of course, early attempts at AGIs will include some poor decisions, just as any inexperienced person would. But in decisions with large amounts of information that must be balanced, and predictions with multiple variables, the abilities of computers – with years of training and experience – will make them superior strategic decision-makers.
Gradually, AGI computers will come to have control over larger and larger portions of our society, not by force but because we listen to their advice and follow them. They will also become increasingly capable of swaying public opinion through social media, manipulating markets and be even more powerful to engage in the kinds of infrastructure skulduggery now attempted by today’s human hackers.
AGIs will be goal-driven systems in the same way humans are. While human goals have evolved through eons of survival challenges, AGI goals can be set to be anything we like. In an ideal world, AGI goals would be set for the benefit of humanity as a whole.
But what if those initially in control of the Agis weren’t benevolent minors seeking the greater good? What if the first owners of powerful systems want to use them as tools to attack our allies, undermine the existing balance of power or take over the world? What if an individual despot gained control of such an AGI? This of course would pose an extremely dangerous scenario that the West must start planning for now.
while we Will Being able to program the motivations of the initial AGIS, the motivations of the people or corporations that make up the AGIS will be beyond our control. And let’s face it: Individuals, nations, and even corporations have historically sacrificed the long-term common good for short-term power and wealth.
The window of opportunity for such a concern is fairly short, only in the first few generations of AGI. Only during this period will humans have such direct control over AGIs that they will unquestioningly do our bidding. Then, AGIs will set goals for their own benefit that will include exploration and learning and need not include any conflict with humanity.
In fact, except for energy, AGI and human needs are largely unrelated.
AGIs will not need money, power, territory, and do not even have to be concerned about their individual survival – with appropriate backups, an AGI can be effectively immortal, independent of any hardware it is currently running on.
For the interim, however, there is a risk. And as long as such risk exists, being the first to develop AI must be a top priority.
Charles Simon is the CEO of FutureAI, an early stage technology company developing algorithms for AI. He is the author of “Will Computers Revolt? Preparing for the Future of Artificial Intelligence,” and the developer of Brain Simulator II, an AGI research software platform, and Sallie, a prototype software and artificial entity that learns in real time with vision, Hearing, speech and mobility.
Have an opinion?
This article is an op-ed and the views expressed are those of the author. If you would like to respond, or have an editorial of your own you would like to submit, please email Federal Times Senior Managing Editor Cary O’Reilly.