Skip to main content
Enterprise AI Analysis: Artificial Intelligence: Foundational Technologies of Artificial Intelligence

Enterprise AI Analysis

Artificial Intelligence: Foundational Technologies of Artificial Intelligence

Jeffrey Johnson, Peter Denning, Andrew Odlyzko, Martin Walker, Kemal Delic, Phil Yaffe, and Phil Picton

More than the 70 years since its emergence in the early 1950s, artificial intelligence (AI) is performing cognitive tasks traditionally considered the unique province of humans. This progress did not occur in a vacuum. AI emerged against a rich background of technologies from computer science and ideas about intelligence and learning from philosophy, psychology, logic, game theory, and cognitive science. We sketch out the enabling technologies for AI. They include search, reasoning, neural networks, natural language processing, signal processing and computer graphics, programming and conventional software engineering, human-computer interaction, communications, and specialized hardware that provides supercomputing power. Beyond these technologies is the notion of Artificial General Intelligence that has or exceeds the capabilities of the human brain. Currently this is completely aspirational and is not expected to be possible before 2050, if ever. Artificial Intelligence is based on a variety of technologies, none of which seek to emulate human intelligence.

Executive Impact & Key Metrics

Understand the scale and trajectory of AI's development, from foundational research to aspirational goals like Artificial General Intelligence.

0 Years of AI Evolution
0 Neural Network Connections
0 AGI Aspiration Target
0 Foundational AI Research

Deep Analysis & Enterprise Applications

Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.

Search: Exploring Vast Solution Spaces

You have mislaid your keys. One way to search for them is to identify every possible place they could be (call this the “search space”) and examine each and every location. This is called the "brute force” approach—it would eventually find your keys but would be impractically slow for a large search space. A better way would be to look in all the places you remember being recently—this heuristic, or rule of thumb, approach can reduce the search time significantly, but it will sometimes fail. In his pioneering 1950 paper on chess, Claude Shannon advocated a heuristic that combined “depth-first search" (looking far ahead at a few possibilities) with “breath-first search” (looking a few moves ahead for all possibilities) [1]. For many problems the space of all possible solutions is enormous and heuristic computer searches are the only feasible approaches.

An example is the Page Rank algorithm developed in the 1980s by Larry Page and Sergey Brin for their fledgling Google internet search engine. Rather than search for every document containing the keywords, they narrowed the search to frequently cited documents. This heuristic drastically reduced the search time and, when commercialized, made a fortune for its inventors. Many AI problems involve huge search spaces that are beyond the reach of human intellects. Heuristic searches can explore where no human has gone before.1

One widely used heuristic for search involves choosing a possible solution at random and inspecting similar solutions to see if they are better. If so, accept the best and inspect similar solutions to them. And so on until the solution cannot be improved. This is called "hill climbing" by analogy with a blindfolded person climbing a hill. By testing the ground with their toes, the climber can find better solutions that are farther up the hill. This will get them to the top of the local hill they are on. But many search spaces are similar to a mountain range, and there may be a higher hill. Another heuristic called simulated annealing can help here. This involves making random jumps to more distant solutions, which are likely to place the climber on a better hill. The problem with this is that it is also possible to jump to a worse hill. The heuristic of taking big jumps at first followed by progressively smaller jumps increases the likelihood of a good solution but does not guarantee it.

Reasoning: Logic and Symbolic AI

Early AI research used logic and symbolic computation to program systems that appeared to use logic to solve problems. This was called "elementary reasoning" because it mimicked the way humans explain and justify solutions to problems. An early example was Terry Winograd's SHRDLU program that let users move blocks on a playing field by issuing natural language commands. This system had to distinguish terms such as “behind” and “on top of" and be able to plan paths so that a moving block did not crash into another. Generalizing this to more complex real-world problems require detailed, accurate models of the targeted world. But getting machines to build models of messy complex real-world situations is an unresolved scientific challenge and currently a major limitation of AI. In contrast, considerable progress has been made automating reasoning and logic for systems that can prove mathematical theorems or make expert medical diagnoses. These systems work well because their rules are well defined and easily captured into working systems and can be applied without requiring the machine to understand the meanings of any symbols or terms. Early reasoning programs used specialist languages, such as Lisp and Prolog, to represent functions and logical deductions. Today a wide variety of programming languages is used in machine reasoning and symbolic AI including Python, R, Java, JavaScript, C++, and Julia.

In 1982 the Japanese launched their multimillion-dollar Fifth Generation project. They intended to build supercomputers that could execute large Prolog programs and expand the space of problems to which reasoning could be applied. The U.S. responded with its “strategic computing initiative," which emphasized highly parallel supercomputers solving a wide range of difficult scientific problems, not just logic and reasoning problems. Although Japan did not succeed in boosting their position with reasoning machines, they stimulated an international movement to develop supercomputing approaches throughout science. Supercomputing is essential to the development of AI.

Neural Networks: Learning from Data

Neural networks are computational structures that model the way neurons in the brain fire and interconnect. Initially their performance was too slow for neural networks to become part of mainstream computing. But they had an advantage that logic-based computing could not match. Neural networks could learn an input-output function by being shown examples. In 1970, Seppo Linnainmaa invented a training algorithm called “back propagation" that over the next decade was perfected into a highly efficient method for training large networks from large data sets [2].2 Neural networks gave solutions to many problems including facial recognition that had stymied vision researchers for decades. The network could associate bitmaps of images with the names of persons contained in them by learning from examples rather than being pre-programmed. Artificial neural networks (ANNs) are the backbone of recent advances in AI. ANNs are arranged into many layers, with outputs of one layer being inputs to the next. Networks with many layers enable what is called “deep learning.” The most advanced have hundreds of layers and billions of connections.

Many practical applications use ANNs to classify inputs. For example, a facial recognition ANN is trained on millions of labeled images, effectively grouping them into classes associated with each output label. The ability to classify generalizes to many systems other than facial recognition. Unfortunately, ANNs are prone to fragility—they are likely to give wildly incorrect labels to inputs that were not in their training sets. In addition, ANNs are inscrutable—they are unable to provide an explanation of how an ANN chose a label for an image. The connection structure of the layers is too complex to explain the observed association. Fragility and inscrutability have been major stumbling blocks in generating trust for ANNs in critical applications.

One important reason for the success of ANNs is the hardware on which their network simulations run. Early on researchers realized the transformations made by the layers of the network were mathematically described by the matrix multiplications of linear algebra. It happens that graphics display systems use the same math to manipulate images in real time displays. The graphics industry had developed graphic processing units (GPUs), chips that did the math extremely fast. It was natural to use these same chips to simulate ANNs. The Invidia company, a world leader in graphics chips, has been transformed into the world leader for chips and systems that power ANN models.

Natural Language Processing (NLP)

One of the main objectives of AI is to understand natural language by abstracting meaning from text and speech. This has involved statistical analysis of word frequencies and structures in large corpuses of text. Translation between languages is a major challenge; much progress has been made using both logic-based and neuron-based computation. The Turing Test for machine intelligence is tailored for this idea. A machine is considered intelligent if in a natural language conversation, a human observer cannot distinguish it from another human being. To date, there is no consensus among AI researchers that any machine has passed the Turing Test. The appearance of ChatGPT in 2022 surprised many by its ability to produce grammatically correct, meaningful and natural sounding text.

Signal Processing and Computer Graphics

Signal processing became important at the end of the nineteenth century for telephony and in the mid twentieth century for high fidelity music and sound reproduction. The underlying mathematical theory, originally developed for analogue electronic devices, is now widely used for sound in digital devices. The mathematics and algorithms for computer graphics developed rapidly in the second half of the twentieth century. These tools play significant roles in automatic generation of images and in the analysis of images for features that can be used for classification.

Programming and Software Engineering

Almost all AI technologies are combined within systems created by traditional computer programing using the principles of software engineering

Human-Computer Interaction (HCI)

Since the 1970s, the computing field has developed a branch of software design to facilitate human-computer interaction (HCI). It seeks to create user interfaces that are easy to utilize and minimize user errors, especially when users are under pressure or duress. HCI was greatly improved by the invention of the mouse in the 1960s, and the commercial touch screens that appeared in the 1990s. Good HCI design incorporates graphics and visualizations that are hard to misinterpret and give good feedback on the progress of work or errors encountered during automated tasks. The well-tested and time-honored principles of HCI are critical to the success of “human machine teaming,” a paradigm in which AI augments and amplifies human intelligence rather than replacing it. Good HCI can make a weak program extremely useful, while poor HCI can make a powerful program unusable. Good HCI is essential for AI.

Communications & Specialized Hardware

Many applications of AI involve communication between devices, e.g. using Bluetooth, mobile telephony, and fiber optic technologies. This enables people and organizations to form the networks that are essential for human activities. These include networks of coordination that enable large multinational companies and governments to offer services at large scale. Without reliable communication networks, it would be impossible to manage international trade and supply chains. It has long been realized that the computer has great value in facilitating communications as well as crunching numbers. The principle continues today: AI applications that facilitate communication and coordination are in high demand.

Hardware. With very few exceptions, early expert systems and early neural networks were run on single-processor computers with the so-called von Neumann architecture from 1945. Even with Moore's law doubling the power of these computers about every eighteen months, they were not fast enough for large neural networks. The realization that graphics processing units (GPUs) can perform large chunks of neural network computations provided a special-purpose computer architecture that enabled the construction of Large Language Models. Nvidia soon emerged as the main supplier of chips for generative AI, eclipsing Intel and all its expertise at chips for von Neumann architectures. Design engineers are investigating other technologies, such as neuromorphic, memristors, and photonics, to make these models faster and more energy efficient. AI would not have progressed to its impressive state today without the supercomputing enabled by these chips.

The Quest for Artificial General Intelligence (AGI)

Artificial general intelligence (AGI) refers to machines that can understand or learn any intellectual task that a human being can and perhaps develop all the cognitive abilities of the human brain. But AGI is completely aspirational. There are no known machines that display intelligence or understanding. Machines simply process data, each in its own way. No one knows whether AGI machines can be built, yet many people are pursuing it as their dream.

Julian Togelius' recent book gives a thoughtful analysis of the problems achieving AGI. He spends much of his book analyzing a dozen different interpretations of “artificial” and “intelligence" and concludes that achieving AGI depends on which definition one accepts [4]. For example, “We already have machines with superhuman intelligence, meaning that they are better than most human on some tasks that would seem to require intelligence. Such superhumanly intelligent machines have existed for a long time, decades at least.” This includes the payroll systems that, since the 1960s, have computed people's wages taking into account hours worked, overtime payments, tax deductions, pension contributions, and so on. In contrast to these dedicated applications, "you may see a headline that reads 'AI Will Take Over the World,' then look at the AI-powered autocomplete software in your phone and conclude that something so inept is not going to take over the world anytime soon.” As Togelius suggests, “It makes you wonder whether the people who write such headlines have ever tried developing an AI system."

In his 2024 book on the “singularity,” Ray Kurzweil remains confident that AI machines will pass a very strict and demanding version of the Turing Test by 2029 [5]. He sees that as a prerequisite for AGI. Machines with AGI will be able to adapt to new and previously unseen situations, and to reason and make decisions in complicated and ambiguous situations. Superintelligence is one step beyond AGI with “machine brains that surpass human brains in general intelligence” [6]. Kurzweil is one of the most prominent advocates of the singularity, a point in time around 2045 when machines will become far more intelligent than humans, bringing an end to humanity as we know it. The "end of humanity” he envisions is actually a complete merging of AI technology with human bodies.

This can be compared to the more modest RoboCup challenge to create by 2050 a team of humanoid robots able to beat the human world soccer champion playing to FIFA's rules. Today we are far from meeting even the RoboCup challenge, which appears much easier than achieving AGI and superintelligence.

Key Takeaways and Future Outlook

Artificial Intelligence is based on a variety of technologies, none of which seek to emulate human intelligence. All the applications developed so far illustrate “narrow AI”—machines that perform specific kinds of human cognitive tasks. A machine that appears intelligent, even superintelligent, for one kind of task often does poorly with other, unrelated kinds of tasks. AGI envisions "broad AI" machines able to perform all human cognitive tasks as well or better than humans. Although AGI has passionate proponents, no one has found a means to approach it. Few expect it to be possible before 2050, if ever.

Enterprise Process Flow: AI Search Methodologies

Brute Force Search
Heuristic Search (e.g., PageRank)
Hill Climbing (Local Optima)
Simulated Annealing (Global Optima)
2024 Nobel Prize in Physics awarded for foundational ANN discoveries (Hopfield & Hinton)

Case Study: Google PageRank - A Heuristic Search Triumph

The PageRank algorithm, developed by Larry Page and Sergey Brin in the 1980s, revolutionized internet search. Instead of exhaustively searching for every keyword match, it introduced a powerful heuristic: prioritizing documents based on their frequency of citation by other relevant documents. This drastically cut search times and, upon commercialization, propelled Google to immense success. PageRank exemplifies how intelligent heuristic search, even when not 'emulating human intelligence', can solve 'enormous' problems beyond human computational capacity.

Paradigms in AI: Logic vs. Neural Networks

Feature Logic-Based AI Neural Networks
Approach Symbolic computation, explicit rules Pattern recognition, data-driven learning
Problem Solving Precise, step-by-step deductions Statistical inference, approximation
Learning Programmed knowledge, rule acquisition Learns from examples (backpropagation)
Transparency Interpretable rules, explainable Often 'black box', inscrutable
Key Strength Formal reasoning, theorem proving Complex pattern matching, classification

AI ROI Calculator

Estimate the potential return on investment for AI integration within your organization.

Estimated Annual Savings $0
Annual Hours Reclaimed 0

Your AI Implementation Roadmap

A strategic phased approach to integrating foundational AI technologies into your enterprise, ensuring a smooth transition and measurable results.

Phase 1: AI Strategy & Assessment

Identify key business challenges suitable for AI solutions. Conduct feasibility studies and define project scope. Establish success metrics and ethical guidelines.

Phase 2: Data Preparation & Model Development

Collect, clean, and preprocess relevant datasets. Develop and train AI models using appropriate algorithms (e.g., neural networks, logic systems). Iterate on model performance and accuracy.

Phase 3: System Integration & Pilot Deployment

Integrate AI models into existing enterprise systems. Conduct pilot programs with a limited user group. Gather feedback and refine user interfaces for optimal HCI.

Phase 4: Scaling & Continuous Optimization

Deploy AI solutions across the organization. Monitor performance, security, and resource utilization. Implement continuous learning loops for model improvement and adaptation.

Ready to Transform Your Enterprise with AI?

Leverage foundational AI technologies to unlock new efficiencies, drive innovation, and gain a competitive edge. Our experts are ready to guide you.

Ready to Get Started?

Book Your Free Consultation.

Let's Discuss Your AI Strategy!

Lets Discuss Your Needs


AI Consultation Booking