acm - an acm publication

Articles

Artificial Intelligence: Foundational Technologies of Artificial Intelligence

Ubiquity, Volume 2025 Issue July, July 2025 | BY Jeffrey Johnson, Peter Denning, Andrew Odlyzko, Martin Walker, Kemal Delic, Phil Yaffe, Phil Picton

|

Full citation in the ACM Digital Library  | PDF


Ubiquity

Volume 2025, Number July (2025), Pages 1-10

Ubiquity Symposium: Artificial Intelligence: Foundational Technologies of Artificial Intelligence
Jeffrey Johnson, Peter Denning, Andrew Odlyzko, Martin Walker, Kemal Delic, Phil Yaffe, Phil Picton
DOI: 10.1145/3747355

More than the 70 years since its emergence in the early 1950s, artificial intelligence (AI) is performing cognitive tasks traditionally considered the unique province of humans. This progress did not occur in a vacuum. AI emerged against a rich background of technologies from computer science and ideas about intelligence and learning from philosophy, psychology, logic, game theory, and cognitive science. We sketch out the enabling technologies for AI. They include search, reasoning, neural networks, natural language processing, signal processing and computer graphics, programming and conventional software engineering, human-computer interaction, communications, and specialized hardware that provides supercomputing power. Beyond these technologies is the notion of Artificial General Intelligence that has or exceeds the capabilities of the human brain. Currently this is completely aspirational and is not expected to be possible before 2025, if ever. Artificial Intelligence is based on a variety of technologies, none of which seek to emulate human intelligence.

More than the 70 years since its emergence as a field of study in the early 1950s, artificial intelligence (AI) has begotten many new technologies for performing cognitive tasks traditionally considered the unique province of humans. This progress did not occur in a vacuum. AI emerged against a rich background of technologies from computer science and ideas about intelligence and learning from philosophy, psychology, logic, game theory, and cognitive science. Our purpose in this essay is to sketch out the enabling technologies for AI. They include:

  • Search; looking for solutions to problems in a space of possible answers
  • Reasoning; applying rules and facts to make logical deductions
  • Neural networks; systems of artificial neurons that learn from data
  • Natural language processing; systems that analyze text and discern its meanings
  • Signal processing and computer graphics
  • Programming and software engineering; almost all AI systems use conventional software engineering to assemble reasoning and machine learning modules into useful systems.
  • Human-computer interaction (HCI); systems designed to interact with people
  • Communications; networked communication including mobile phones and the internet
  • Hardware; highly parallel chips that bring supercomputing power to the technologies referred to above

Each of these domains is rich in its own complexities. We have, however, found a general awareness of these technologies is usually sufficient to understand how AI services and products work—and therefore to evaluate their weaknesses and strengths.

A Sketch of the Foundational Technologies

Search. You have mislaid your keys. One way to search for them is to identify every possible place they could be (call this the "search space") and examine each and every location. This is called the "brute force"approach—it would eventually find your keys but would be impractically slow for a large search space. A better way would be to look in all the places you remember being recently—this heuristic, or rule of thumb, approach can reduce the search time significantly, but it will sometimes fail. In his pioneering 1950 paper on chess, Claude Shannon advocated a heuristic that combined "depth-first search" (looking far ahead at a few possibilities) with "breath-first search" (looking a few moves ahead for all possibilities) [1]. For many problems the space of all possible solutions is enormous and heuristic computer searches are the only feasible approaches.

An example is the Page Rank algorithm developed in the 1980s by Larry Page and Sergey Brin for their fledgling Google internet search engine. Rather than search for every document containing the keywords, they narrowed the search to frequently cited documents. This heuristic drastically reduced the search time and, when commercialized, made a fortune for its inventors. Many AI problems involve huge search spaces that are beyond the reach of human intellects. Heuristic searches can explore where no human has gone before.1

One widely used heuristic for search involves choosing a possible solution at random and inspecting similar solutions to see if they are better. If so, accept the best and inspect similar solutions to them. And so on until the solution cannot be improved. This is called "hill climbing" by analogy with a blindfolded person climbing a hill. By testing the ground with their toes, the climber can find better solutions that are farther up the hill. This will get them to the top of the local hill they are on. But many search spaces are similar to a mountain range, and there may be a higher hill. Another heuristic called simulated annealing can help here. This involves making random jumps to more distant solutions, which are likely to place the climber on a better hill. The problem with this is that it is also possible to jump to a worse hill. The heuristic of taking big jumps at first followed by progressively smaller jumps increases the likelihood of a good solution but does not guarantee it.

Genetic algorithms provide another heuristic for exploring large search spaces. This involves representing possible solutions as strings of numbers analogous to chromosomes. The search process begins by creating a population of chromosomes and successively combining parts of chromosomes with parts of others. The best (or 'fittest') solutions are retained for further iterations of the combination process. This approach corresponds to the combination of genes in biological reproduction. In principle the best parts of two mediocre solutions may be combined to create a much better solution. To ensure that possibilities outside the original random population are included, a process analogous to mutation is introduced that randomly changes some of the numbers.

Reasoning. Early AI research used logic and symbolic computation to program systems that appeared to use logic to solve problems. This was called "elementary reasoning" because it mimicked the way humans explain and justify solutions to problems. An early example was Terry Winograd's SHRDLU program that let users move blocks on a playing field by issuing natural language commands. This system had to distinguish terms such as "behind" and "on top of" and be able to plan paths so that a moving block did not crash into another. Generalizing this to more complex real-world problems require detailed, accurate models of the targeted world. But getting machines to build models of messy complex real-world situations is an unresolved scientific challenge and currently a major limitation of AI. In contrast, considerable progress has been made automating reasoning and logic for systems that can prove mathematical theorems or make expert medical diagnoses. These systems work well because their rules are well defined and easily captured into working systems and can be applied without requiring the machine to understand the meanings of any symbols or terms. Early reasoning programs used specialist languages, such as Lisp and Prolog, to represent functions and logical deductions. Today a wide variety of programming languages is used in machine reasoning and symbolic AI including Python, R, Java, JavaScript, C++, and Julia.

In 1982 the Japanese launched their multimillion-dollar Fifth Generation project. They intended to build supercomputers that could execute large Prolog programs and expand the space of problems to which reasoning could be applied. The U.S. responded with its "strategic computing initiative," which emphasized highly parallel supercomputers solving a wide range of difficult scientific problems, not just logic and reasoning problems. Although Japan did not succeed in boosting their position with reasoning machines, they stimulated an international movement to develop supercomputing approaches throughout science. Supercomputing is essential to the development of AI.

Neural networks. Neural networks are computational structures that model the way neurons in the brain fire and interconnect. Initially their performance was too slow for neural networks to become part of mainstream computing. But they had an advantage that logic-based computing could not match. Neural networks could learn an input-output function by being shown examples. In 1970, Seppo Linnainmaa invented a training algorithm called "back propagation" that over the next decade was perfected into a highly efficient method for training large networks from large data sets [2].2 Neural networks gave solutions to many problems including facial recognition that had stymied vision researchers for decades. The network could associate bitmaps of images with the names of persons contained in them by learning from examples rather than being pre-programmed. Artificial neural networks (ANNs) are the backbone of recent advances in AI. ANNs are arranged into many layers, with outputs of one layer being inputs to the next. Networks with many layers enable what is called "deep learning." The most advanced have hundreds of layers and billions of connections.

Many practical applications use ANNs to classify inputs. For example, a facial recognition ANN is trained on millions of labeled images, effectively grouping them into classes associated with each output label. The ability to classify generalizes to many systems other than facial recognition. Unfortunately, ANNs are prone to fragility—they are likely to give wildly incorrect labels to inputs that were not in their training sets. In addition, ANNs are inscrutable—they are unable to provide an explanation of how an ANN chose a label for an image. The connection structure of the layers is too complex to explain the observed association. Fragility and inscrutability have been major stumbling blocks in generating trust for ANNs in critical applications.

One important reason for the success of ANNs is the hardware on which their network simulations run. Early on researchers realized the transformations made by the layers of the network were mathematically described by the matrix multiplications of linear algebra. It happens that graphics display systems use the same math to manipulate images in real time displays. The graphics industry had developed graphic processing units (GPUs), chips that did the math extremely fast. It was natural to use these same chips to simulate ANNs. The Invidia company, a world leader in graphics chips, has been transformed into the world leader for chips and systems that power ANN models.

Natural language processing. One of the main objectives of AI is to understand natural language by abstracting meaning from text and speech. This has involved statistical analysis of word frequencies and structures in large corpuses of text. Translation between languages is a major challenge; much progress has been made using both logic-based and neuron-based computation. The Turing Test for machine intelligence is tailored for this idea. A machine is considered intelligent if in a natural language conversation, a human observer cannot distinguish it from another human being. To date, there is no consensus among AI researchers that any machine has passed the Turing Test. The appearance of ChatGPT in 2022 surprised many by its ability to produce grammatically correct, meaningful and natural sounding text.

Signal processing and computer graphics. Signal processing became important at the end of the nineteenth century for telephony and in the mid twentieth century for high fidelity music and sound reproduction. The underlying mathematical theory, originally developed for analogue electronic devices, is now widely used for sound in digital devices. The mathematics and algorithms for computer graphics developed rapidly in the second half of the twentieth century. These tools play significant roles in automatic generation of images and in the analysis of images for features that can be used for classification.

Programming and software engineering. Almost all AI technologies are combined within systems created by traditional computer programing using the principles of software engineering

Human-computer interaction. Since the 1970s, the computing field has developed a branch of software design to facilitate human-computer interaction (HCI). It seeks to create user interfaces that are easy to utilize and minimize user errors, especially when users are under pressure or duress. HCI was greatly improved by the invention of the mouse in the 1960s, and the commercial touch screens that appeared in the 1990s. Good HCI design incorporates graphics and visualizations that are hard to misinterpret and give good feedback on the progress of work or errors encountered during automated tasks. The well-tested and time-honored principles of HCI are critical to the success of "human machine teaming," a paradigm in which AI augments and amplifies human intelligence rather than replacing it. Good HCI can make a weak program extremely useful, while poor HCI can make a powerful program unusable. Good HCI is essential for AI.

Communications. Many applications of AI involve communication between devices, e.g. using Bluetooth, mobile telephony, and fiber optic technologies. This enables people and organizations to form the networks that are essential for human activities. These include networks of coordination that enable large multinational companies and governments to offer services at large scale. Without reliable communication networks, it would be impossible to manage international trade and supply chains. It has long been realized that the computer has great value in facilitating communications as well as crunching numbers. The principle continues today: AI applications that facilitate communication and coordination are in high demand.

Hardware. With very few exceptions, early expert systems and early neural networks were run on single-processor computers with the so-called von Neumann architecture from 1945. Even with Moore's law doubling the power of these computers about every eighteen months, they were not fast enough for large neural networks. The realization that graphics processing units (GPUs) can perform large chunks of neural network computations provided a special-purpose computer architecture that enabled the construction of Large Language Models. Nvidia soon emerged as the main supplier of chips for generative AI, eclipsing Intel and all its expertise at chips for von Neumann architectures. Design engineers are investigating other technologies, such as neuromorphic, memristors, and photonics, to make these models faster and more energy efficient. AI would not have progressed to its impressive state today without the supercomputing enabled by these chips.

The Quest for Artificial General Intelligence

Artificial general intelligence (AGI) refers to machines that can understand or learn any intellectual task that a human being can and perhaps develop all the cognitive abilities of the human brain. But AGI is completely aspirational. There are no known machines that display intelligence or understanding. Machines simply process data, each in its own way. No one knows whether AGI machines can be built, yet many people are pursuing it as their dream.

Julian Togelius' recent book gives a thoughtful analysis of the problems achieving AGI. He spends much of his book analyzing a dozen different interpretations of "artificial" and "intelligence" and concludes that achieving AGI depends on which definition one accepts [4]. For example, "We already have machines with superhuman intelligence, meaning that they are better than most human on some tasks that would seem to require intelligence. Such superhumanly intelligent machines have existed for a long time, decades at least." This includes the payroll systems that, since the 1960s, have computed people's wages taking into account hours worked, overtime payments, tax deductions, pension contributions, and so on. In contrast to these dedicated applications, "you may see a headline that reads 'AI Will Take Over the World,' then look at the AI-powered autocomplete software in your phone and conclude that something so inept is not going to take over the world anytime soon." As Togelius suggests, "It makes you wonder whether the people who write such headlines have ever tried developing an AI system."

In his 2024 book on the "singularity," Ray Kurzweil remains confident that AI machines will pass a very strict and demanding version of the Turing Test by 2029 [5]. He sees that as a prerequisite for AGI. Machines with AGI will be able to adapt to new and previously unseen situations, and to reason and make decisions in complicated and ambiguous situations. Superintelligence is one step beyond AGI with "machine brains that surpass human brains in general intelligence" [6]. Kurzweil is one of the most prominent advocates of the singularity, a point in time around 2045 when machines will become far more intelligent than humans, bringing an end to humanity as we know it. The "end of humanity" he envisions is actually a complete merging of AI technology with human bodies.

This can be compared to the more modest RoboCup challenge to create by 2050 a team of humanoid robots able to beat the human world soccer champion playing to FIFA's rules. Today we are far from meeting even the RoboCup challenge, which appears much easier than achieving AGI and superintelligence.

Conclusions

Artificial Intelligence is based on a variety of technologies, none of which seek to emulate human intelligence. All the applications developed so far illustrate "narrow AI"—machines that perform specific kinds of human cognitive tasks. A machine that appears intelligent, even superintelligent, for one kind of task often does poorly with other, unrelated kinds of tasks. AGI envisions "broad AI" machines able to perform all human cognitive tasks as well or better than humans. Although AGI has passionate proponents, no one has found a means to approach it. Few expect it to be possible before 2050, if ever.

References

[1] Shannon, C. E. XXII. Programming a computer for playing chess. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science 41, 314 (1950), 256–275. Reprinted in: Levy, D. (ed.) Computer Chess Compendium, Springer, 1988.

[2] Rumelhart, D., Hinton, G., and Williams, R. Learning representations by back-propagating errors. Nature 323, 6088 (198), 533–536.

[3] Benjamin Weiser, 'Here's What Happens When Your Lawyer Uses ChatGPT', New York Times, May 27, 2023 https://www.nytimes.com/2023/05/27/nyregion/avianca-airline-lawsuit-chatgpt.html?smid=tw-nytimes&smtyp=cur

[4] Togelius, J. Artificial General Intelligence. MIT Press, 2024.

[5] Kurzweil, R. The Singularity is Nearer: When we Merge with AI. Viking, 2024.

[6] Bostrom, N. Superintelligence: Paths, Dangers, Strategies. Oxford University Press, 2014.

Authors

Jeffrey Johnson is Professor of Complexity Science and Design at Open University. He has published numerous articles and books on complex systems science, artificial intelligence, mathematics, planning, and design. He is CEO of Vision Scientific Ltd and AI-Inc Ltd and has led many academic and commercial projects.

Peter J. Denning is a Distinguished Professor of computer science at the Naval Postgraduate School in Monterey, California. He is a past president of ACM (1980-82). He received the IEEE Computer Pioneer Award in 2021. His most recent books are Computational Thinking (with Matti Tedre, MIT Press, 2019) and Navigating a Restless Sea (with Todd Lyons, Waterside, 2024).

Andrew Odlyzko has had a long career in research and research management at Bell Labs, AT&T Labs, and most recently at the University of Minnesota, where he built an interdisciplinary research center, and is now a professor in the School of Mathematics. He has worked in a number of areas of mathematics, computing, and communications, and is concentrating on diffusion of technological innovation. His recent works are available through his web page. is a Distinguished Professor of computer science at the Naval Postgraduate School in Monterey, California. He is a past president of ACM (1980-82). He received the IEEE Computer Pioneer Award in 2021. His most recent book is Computational Thinking (with Matti Tedre, MIT Press, 2019).

Kemal Delic is a senior visiting research fellow at the Center for Complexity and Design at the Open University. He is the co-founder of AI-Inc, Ltd. and a lecturer at the University of Grenoble and University of Sarajevo. He is advisor and expert evaluator to the European Commission. He previously held positions as a senior enterprise architect and senior technologist and scientist at Hewlett-Packard. He can be found on Twitter @OneDelic.

Martin Antony Walker's (Ph.D. – 1969, supervisor Sir Roger Penrose) career has spanned research in mathematical physics, and high-performance computing product research, development, and marketing with several leading high-tech firms in Canada, the USA, and Europe. He has evaluated EC research, advised UNESCO programs, and operated a scientific computing consultancy prior to retirement. His current interest is in applications of AI to science.

Philip Yaffe is a former reporter/feature writer with The Wall Street Journal, has more than 40 years' experience as a technical/industrial communication consultant. He is a founding member of the first bilingual Toastmasters International club in Brussels, Belgium (there are now nine of them). He has a special interest in communicating the essence and romance of science to people who think they hate it.

Phil Picton is an Emeritus Professor at the Open University. He has published numerous articles and books, mainly on the subject of the application of artificial intelligence to engineering problems.

Footnotes

1. By this we mean that machines can enable humans to explore huge spaces of possibilities. This practical amplification of human intelligence has been a reality for many years. Whether or not machines can initiate searches to amplify their own intelligence is an unresolved question.

2. In October 2024 the Royal Swedish Academy of Sciences awarded the Nobel Prize in Physics to John J. Hopfield of Princeton University, NJ, USA and Geoffrey E. Hinton of the University of Toronto, Canada "for foundational discoveries and inventions that enable machine learning with artificial neural networks."

2025 Copyright held by the Owner/Author.

The Digital Library is published by the Association for Computing Machinery. Copyright © 2025 ACM, Inc.

COMMENTS

POST A COMMENT
Leave this field empty