Perceptrons: A Brain in a Function
Learn about the inspiration behind what is now the field of deep learning and generative AI.
We'll cover the following...
The simplest neural network architecture—the perceptron was inspired by biological research to understand the basis of mental processing in an attempt to represent the function of the brain with mathematical formulae.
From tissues to TLUs
The recent popularity of AI algorithms might give the false impression that this field is new. Many recent models are based on discoveries made decades ago that have been reinvigorated by the massive computational resources available in the cloud and customized hardware for parallel matrix computations, such as Graphical Processing Units (GPUs), Tensor Processing Units (TPUs), and Field-Programmable Gate Array (FPGAs).
If we consider research on neural networks to include their biological inspiration as well as computational theory, this field is over a hundred years old. Indeed, one of the first neural networks described appears in the detailed anatomical illustrations of 19th-century scientist
This observation of simple neuronal cells interconnected in large networks led computational researchers to hypothesize how mental activity might be represented by simple, logical operations that, combined, yield complex mental phenomena. The original “automata theory” is usually traced to a 1943 article by Warren McCulloch and Walter Pitts of the Massachusetts Institute of
where
Visually and conceptually, there is some similarity between McCulloch and Pitts’ model and the biological neuron that inspired it. Their model integrates inputs into an output signal, just as the natural dendrites (short, input “arms” of the neuron that receive signals from other cells) of a neuron synthesize inputs into a single output via the axon (the long “tail” of the cell, which passes signals received from the dendrites along to other neurons). We might imagine that, just as neuronal cells are composed into networks to yield complex biological circuits, these simple units might be connected to simulate sophisticated decision processes.
Indeed, using this simple model, we can already start to represent several logical operations. If we consider a simple case of a neuron with one input, we can see that a TLU can solve an identity or negation function.
For an identity operation that simply returns the input as output, the weight matrix would have
TLU Logic for Identity Operations
Identity | |
Input | Output |
1 | 1 |
0 | 0 |
Similarly, for a negation operation, the weight matrix could be a negative identity matrix, with a threshold at