paint-brush
The Cognitive Paradigm: Exploring Brain-Inspired AI Development by@edemgold
1,452 reads
1,452 reads

The Cognitive Paradigm: Exploring Brain-Inspired AI Development

by Edem GoldMay 10th, 2023
Read on Terminal Reader
Read this story w/o Javascript

Too Long; Didn't Read

In this article, we will attempt to understand the brain-inspired approach to building AI systems. We will discuss how scientists Norman Weiner brought about the convergence of neuroscience and computer science. Frank Rosenblatt's Perceptron was the first real attempt to mimic human intelligence and how its failure brought about ground-breaking work which would serve as the platform for Neural Networks.

People Mentioned

Mention Thumbnail
featured image - The Cognitive Paradigm: Exploring Brain-Inspired AI Development
Edem Gold HackerNoon profile picture


"Our Intelligence is what makes us human and AI is an extension of that Quality" -Yan LeCun


Since the advent of Neural Networks (also known as artificial neural networks), the AI industry has enjoyed unparalleled success. Neural Networks are the driving force behind modern AI systems and they are modeled after the human brain. Modern AI research involves creating and implementing algorithms that aim to mimic the neural processes of the human brain to create systems that learn and act in ways similar to human beings.


In this article, we will attempt to understand the brain-inspired approach to building AI systems. It is worth noting that while this topic is an inherently broad one, I will be as brief and succinct as possible. I plan to treat sub-topics that have more intricate sub-branches as standalone articles, and I will, of course, leave references at the end of the article.


Content breakdown

How I hope to approach this:

  • History of the brain-inspired approach to AI: Here we'll discuss how scientists Norman Weiner ann Warren McCulloch brought about the convergence of neuroscience and computer science, how Frank Rosenblatt's Perceptron was the first real attempt to mimic human intelligence, and how its failure brought about ground-breaking work which would serve as the platform for Neural Networks.


  • How the human brain works and how it relates to AI systems: In this section, we'll dive into the biological basis for the brain-inspired approach to AI. We will discuss the basic structure and functions of the human brain, understand its core building block, the neuron, and how they work together to process information and enable complex actions.


  • The Core Principles behind the Brain-inspired Approach to AI: Here we will discuss the fundamental concepts behind the brain-inspired approach to AI. We will explain how concepts such as; Neural networks, Hierarchical processing, and plasticity and how techniques of parallel processing, distributed representations, and recurrent feedback aid AI in mimicking the brain's functioning.

  • Challenges in building AI systems modeled after the human brain: Here we will talk about the challenges and limitations inherent in attempting to build systems that mimic the human brain. Challenges such as; the complexity of the brain, and the lack of a unified theory of cognition explore the ways these challenges and limitations are being addressed.


Let us begin!


The History of the brain-inspired Approach to AI

The drive to build machines that are capable of intelligent behavior owes much of its inspiration to MIT Profesor, Norbert Weiner. Norbert Weiner was a child prodigy who could read by the age of three. He had a broad knowledge of various fields such as Mathematics, Neurophysiology, medicine, and physics.


Norbert Weiner believed that the main opportunities in science lay in exploring what he termed as Boundary Regions -areas of study that are not clearly within a certain discipline but rather a mixture of disciplines like the study of medicine and engineering coming together to create the field of Medical Engineering-, he was quoted saying:


"If the difficulty of a physiological problem is mathematical in nature, ten physiologists ignorant of mathematics will get precisely as far as one physiologist ignorant of mathematics"


In the year 1934, Weiner and a couple of other academics gathered monthly to discuss papers involving boundary region science.



It was a perfect catharsis for half-baked ideas, insufficient self-criticism, exaggerated self-confidence and pomposity - Norman Weiner


From these sessions and from his own personal research, Weiner learned about new research on biological nervous systems as well as about pioneering work on electronic computers, and his natural inclination was to blend these two fields. And so, a relationship between neuroscience and computer science was formed. This relationship became the cornerstone for the creation of artificial intelligence, as we know it.


After World War II, Wiener began forming theories about intelligence in both humans and machines and this new field was named Cybernetics. Wiener’s foray into Cybernetics successfully got scientists talking about the possibility of biology fusing with engineering. One of the said scientists was a neurophysiologist named Warren McCulloch, who then proceeded to drop out of Haverford University to study philosophy and psychology at Yale.


While attending a scientific conference in New York, he came in contact with papers written by colleagues on biological feedback mechanisms. The following year, in collaboration with his brilliant 18-year-old protege named Walter Pitts, McCulloch proposed a theory about how the brain works -a theory that would help foster the widespread perception that computers and brains function essentially in the same way.


They based their conclusions on research by McCulloch on the possibility of neurons processing Binary Numbers (for the unknowledgeable, computers communicate via binary numbers). This theory became the foundation for what became the first model of an artificial neural network, which was named the McCulloch-Pitts Neuron(MCP).


The MCP served as the foundation for the creation of the first-ever neural network which came to be known as the perceptron. The Perceptron was created by Psychologist, Frank Rosenblatt who, inspired by the synapses in the brain, decided that as the human brain could process and classify information through synapses(communication between neurons) then perhaps a digital computer could do the same via a neural network.


The Perceptron essentially scaled the MCP neuron from one artificial neuron into a network of neurons, but, unfortunately, the perceptron had some technical challenges which limited its practical application, most notable of its limitations was its inability to perform complex operations(like classifying between more than one item, for example, a perceptron could not perform classification between a cat, a dog, and a bird).


In the year 1969, a book published by Marvin Minsky and Seymour Papert titled Perceptron lay out in detail the flaws of the Perceptron and due to that, research on Artificial Neural Networks stagnated until the proposal of Back Propagation by Paul Werbos.


Back Propagation hopes to solve the issue of classifying complex data that hindered the industrial application of Neural Networks at the time. It was inspired by synaptic plasticity; the way the brain modifies the strengths of connections between neurons and as such improves performance. Back Propagation was designed to mimic the process in the brain that strengthens connections between neurons via a process called weight adjustment.


Despite the early proposal by Paul Werbos, the concept of backpropagation only gained widespread adoption when researchers such as David RumelheartGeoffrey Hinton, and Ronald Williams published papers that demonstrated the effectiveness of backpropagation for training neural networks. The implementation of backpropagation led to the creation of Deep Learning which powers most of the AI systems available in the world.


"People are smarter than today's computers because the brain employs a basic computational architecture that is more suited to deal with a central aspect of the natural information processing tasks that people are so good at." - Parallel Distributed Processing



How the human brain works and how it relates to AI systems



We have discussed how researchers began to model AI to mimic the human brain, let us now look at how the brain works and define the relationship between the brain and AI systems.

How the brain works: A simplified description

The human brain, essentially processes thoughts via the use of neurons, a neuron is made up of 3 core sections; The Dendrite, Axon, and the Soma. The Dendrite is responsible for receiving signals from other Neurons, The Soma processes information received from the Dendrite, and the Axon is responsible for transferring the processed information to the next Dendrite in the sequence.


To grasp how the brain processes thought, imagine you see a car coming towards you, your eyes immediately send electrical signals to your brain through the optical nerve and then the brain forms a chain of neurons to make sense of the incoming signal. The first neuron in the chain collects the signal through its Dendrites and sends it to the Soma to process the signal after the Soma finishes with its task it sends the signal to the Axon which then sends it to the Dendrite of the next neuron in the chain, the connection between Axons and Dendrites when passing on information is called a Synapse. The entire process continues until the brain finds a Sapiotemporal Synaptic Input(that's scientific lingo for; the brain continues processing until it finds an optimal response to the signal sent to it) and then it sends signals to the necessary effectors eg, your legs and then the brain sends a signal to your legs to run away from the oncoming car.


The relationship between the brain and AI systems

The relationship between the brain and AI is largely mutually beneficial with the brain being the main source of inspiration behind the design of AI systems and advances in AI leading to a better understanding of the brain and how it works.


There is a reciprocal exchange of knowledge and ideas when it comes to the brain and AI, and there are several examples that attest to the positively symbiotic nature of this relationship:


  • Neural Networks: Arguably the most significant impact made by the human brain to the field of Artificial Intelligence is the creation of Neural Networks. In essence, Neural Networks are computational models that mimic the function and structure of biological neurons, the architecture of neural networks and their learning algorithms are largely inspired by the way neurons in the brain interact and adapt.


  • Brain Simulations: AI systems have been used to simulate the human brain and study its interactions with the physical world. For example, researchers have Machine Learning techniques to simulate the activity of biological neurons involved in visual processing, and the result has provided insight into how the brain handles visual information.


  • Insights into the brain: Researchers have begun using Machine Learning Algorithms to analyze and gain insights from brain data, and fMRI scans. These insights serve to identify patterns and relationships which would otherwise have remained hidden. The insights gotten can help in the understanding of internal cognitive functions, memory, and decision-making, it also aids in the treatment of brain-native illnesses such as Alzheimer's.


Core Principles Behind the brain-inspired Approach to AI

Here we will discuss several concepts which aid AI in imitating the way the human brain functions. These concepts have helped AI researchers create more powerful and intelligent systems which are capable of performing complex tasks.

Neural Networks

As discussed earlier, neural networks are arguably the most significant impact made by the human brain to the field of Artificial Intelligence. In essence, Neural Networks are computational models that mimic the function and structure of biological neurons, the networks are made up of various layers of interconnected nodes, called artificial neurons, which aid in the processing and transmitting of information, similar to what is done by dendrites, somas, and axons in biological neural networks. Neural Networks are architected to learn from past experiences the same way the brain does.

Distributed Representations

Distributed representations are simply a way of encoding concepts or ideas in a neural network as a pattern along several nodes in a network in order to form a pattern. For example, the concept of smoking could be represented(encoded) using a certain set of nodes in a neural network and so if that network comes along an image of a man smoking it then uses those selected nodes to make sense of the image (it's a lot more complex than that but for the sake of simplicity), this technique aids AI systems in remembering complex concepts or relationships between concepts the same way the brain recognizes and remembers complex stimuli.

Recurrent Feedback

This is a technique used in training AI models where the output of a neural network is returned as input to allow the network to integrate its output as extra data input in training. This is similar to how the brain makes use of feedback loops in order to adjust its model based on previous experiences.

Parallel Processing

Parallel processing involves breaking up complex computational tasks into smaller bits in an effort to process the smaller bits on another processor in an attempt to improve speed. This approach enables AI systems to process more input data faster, similar to how the brain is able to perform different tasks at the same time(multi-tasking).

Attention Mechanisms

This is a technique used which enables AI models to focus on specific parts of input data, it is commonly employed in sectors such as Natural Language Processing which contains complex and cumbersome data. It is inspired by the brain's ability to attend to only specific parts of a largely distracting environment; like your ability to tune into and interact in one conversation out of a cacophony of conversations.

Reinforcement Learning

Reinforcement Learning is a technique used to train AI systems, it was inspired by how human beings learn skills through trial and error. It involves an AI agent receiving rewards or punishments based on its actions, this enables the agent to learn from its mistakes and be more efficient in its future actions (this technique is usually used in the creation of games).

Unsupervised Learning

The brain is constantly receiving new streams of data in the form of sounds, visual content, sensory feelings to the skin, etc and it has to make sense of it all and attempt to form a coherent and logical understanding of how all these seemingly disparate events affect its physical state.

Take this analogy as an example, you feel water drop on your skin, you hear the sound of water droplets dropping quickly on rooftops, you feel your clothes getting heavy and in that instant, you know rain is falling, you then search your memory bank to ascertain if you carried an umbrella and if you did, you are fine, else, you check to see the distance from your current location to your home if it is close, you are fine, else you try to gauge how intense the rain is going to become if it is a light drizzle you can attempt to continue the journey back to your home, but if it is priming to become a shower, then you have to find shelter.


The ability to make sense of seemingly disparate data points(water, sound, feeling, distance) is implemented in Artificial intelligence in the form of a technique called Unsupervised Learning. It is an AI training technique where AI systems are taught to make sense of raw, unstructured data without explicit labeling ( no one tells you rain is falling when it is falling, do they/).

Challenges in Building Brain-Inspired AI Systems

We have spoken about how the approach for using the brain as inspiration for AI systems came about, how the brain relates to AI, and the core principles behind brain-inspired AI. In this section, we are going to talk about some of the technical and conceptual challenges inherent in building AI systems modeled after the human brain.

Complexity

This is a pretty daunting challenge. The brain-inspired approach to AI is based on modeling the brain and building AI systems after that model, but the human brain is an inherently complex system with 100 billion neurons and approximately 600 trillion synaptic connections(each neuron, has on average, 10,000 synaptic connections with other neurons), and these synapses are constantly interacting in dynamic and unpredictable ways. Building AI systems that are aimed to mimic, and hopefully exceed, that complexity is in itself a challenge and requires equally complex statistical models.

Data Requirements for Training Large Models

Open AI's GPT 4, which is, at the moment, the cutting edge of text-based AI models requires 47 GigaBytes of data, in comparison, its predecessor GPT3 was trained on 17 Gigabytes of data, which is approximately 3 orders of magnitude lower, imagine how much GPT 5 will be trained on.


As has been proven in order to get acceptible results, Brain-Inspired AI systems require vast amounts of data and data for tasks especially auditory and visual tasks and this places a lot of emphasis on the creation of data collection pipelines, for instance, Tesla has 780 million miles of driving data and its data collection pipeline adds another million every 10 hours.

Energy Efficiency

Building brain-inspired AI systems that emulate the brain's energy efficiency is a huge challenge. the human brain consumes approximately 20 watts of power, in comparison, Tesla's Autopilot, on specialized chips, consumes about 2,500 watts per second and it takes around 7.5-megawatt hours(MWh) to train an AI model the size of ChatGPT.


The Explainaibility Problem

Developing brain-inspired AI systems that can be trusted by users is crucial to the growth and adoption of AI, but therein lies the problem, the brain, which AI systems are meant to be modeled after, is essentially a black box. The inner workings of the brain are not easy to understand, this is due to the lack of information surrounding how the brain processes thought. There is no lack of research on how the biological structure of the human brain but there is a certain lack of empirical information on the functional qualities of the brain, that is, how thought is formed, how deja vu occurs, etc, and this leads to a problem in the building of brain-inspired AI systems.

The Interdisciplinary Requirements

The act of building brain-inspired AI systems requires the knowledge of experts in different fields, such as; Neuroscience, Computer Science, Engineering, Philosophy, and Psychology. But there lies a challenge there both logistical and foundational, in the sense that getting experts from different fields very financially tasking, and also there lies the problem of knowledge conflict; it is really difficult to get an engineer to care about the psychological effects of what he/she is building, not to talk of the problem of colliding egos.

Summary

In Conclusion, while the brain-inspired approach is the obvious route to building AI systems(we have discussed why), it is wrought with challenges but we can look to the future with the hope that efforts are being made to solve these problems.


Also published here.


References