John Holland received the following accolade from the president of the Santa Fe Institute (SFI), as he recognized his main contribution in terms of the scientific currency, as reported in Holland’s obituary“Obituary: John Henry Holland | the University Record.” n.d. Record.umich.edu. Accessed September 7, 2023. https://record.umich.edu/articles/obituary-john-henry-holland/.:
John (was) rather unique in that he took ideas from evolutionary biology in order to transform search and optimization in computer science, and then he took what he discovered in computer science and allowed us to rethink evolutionary dynamics. This kind of rigorous translation between two communities of thought is a characteristic of very deep minds. And John’s ideas at the interface of the disciplines continue to have a lasting impact on the culture and research of SFI. (Simon 2015)
This obituary note also highlights the importance of the two-way relationship between our computational models and scientific theories. This fruit cannot be reaped if we bypass the process and focus only on algorithms and technology (or the last two components of the process.)
Geoffrey Hinton#
One last example is that of Geoffrey Hinton, who is almost single-handedly responsible for rescuing machine learning from its first winter, by an ingenious idea of applying a chain-rule while backpropagating the redistribution of weights through the hidden layers with respect to the error on the output layer. Without Hinton’s 1986 contribution, known as the famous backpropagation, modern deep-learning neural networks, including Transformer based GPTs wouldn’t be possible. But the note on which Hinton’s seminal paper“Backprop Old.” n.d. https://www.iro.umontreal.ca/~vincentp/ift3395/lectures/backprop_old.pdf. ended needs to be highlighted in the context of this blog:
The learning procedure, in its current form, is not a plausible model of learning in brains. However, applying the procedure to various tasks shows that interesting internal representations can be constructed by gradient descent weight-space, and this suggests that it is worth looking for more biologically plausible ways of doing gradient descent in neural networks. (Rumelhart, Hinton, and Williams 1986, 536)
This only goes to show the culture within which most of the development within computer science was taking place up until the 1990s, that important publications had to put up a disclaimer if their computational models diverged a little from the theories and the scientific process.
In conclusion#
Perhaps Chomsky has the intuitive insight, being a scientist par excellence, that despite so much investment of computing power and training data, GPTs are fascinating as a chatbot and Q/A technology, but they really do not inform us about how intelligence or language works. Perhaps, it is the bypassing of the scientific process that results in the hallucinations, both by the GPTs as well their creators.Klein, Naomi. 2023. “AI Machines Aren’t ‘Hallucinating’. But Their Makers Are.” The Guardian, May 8, 2023, sec. Opinion. https://www.theguardian.com/commentisfree/2023/may/08/ai-machines-hallucinating-naomi-klein.
If you enjoyed reading this blog, you can check our other blogs from this series or the foundational course on neural networks: