Thrilled today to have anticipated around two years prior a paper that DeepMind just released. This is validation that my approach to general intelligence. Here’s the said paper:
Quoted from the paper “Our definition of a symbol draws on the work of the philosopher Charles Sanders Peirce. Peirce outlined three categories of relation — icons, indices, and symbols — whose definitions illuminate the role of convention in establishing meaning.” The authors conclude that “The views we’ve presented here seem to run counter to the past and current AI zeitgeist surrounding symbols, which emphasizes rigid mechanisms and particular representational forms.”
I’ve been a big proponent of C.S. Peirce formulation of evolutionary logic. When one first encounters Peirce, one finds his formulation of signs where he defines icons, indices, and symbols. The DeepMind paper is inspired by their understanding of this slice of Peirce’s work. However, Peirce semiotics is actually much richer than the pedestrian understanding of semiotics.
It’s relieving to find confirmation that one’s approach in fact runs ‘counter to past and current AI’. This may explain why I’ve my approach has difficulty gaining traction in the mainstream AI community. This is also a blessing because to separate oneself from the crowd, one must not be running with the crowd.
The key points of this paper are what the authors describe as symbolic fluency: receptive, constructive, embedded, malleable, separable, meaningful, and graded. Allow me the liberty to explore these 7 characteristics of symbolic fluency so as to mine newer insights.
Let me first point out that the authors allude to an enactivist approach to cognition. An enactivist approach implies that an agent learns about this world by being embodied, embedded, extended and enactive with its environment.
- Embodied involving more than the brain, including a more general involvement of bodily structures and processes.
- Embedded functioning only in a related external environment.
- Enacted involving not only neural processes but also things an organism does.
- Extended into the organism’s environment.
24 years ago, Horst Hendriks-Jansen wrote a book on AI “Catching Ourselves in the Act” where he describes an enactivist approach of AI revolving around 3 core ideas: situated context, interactive emergence, and evolution. It’s very rare that new ideas originate from a complete vacuum. Peirce semiotics is over 100 years old and there have been many deep thinkers who have evolved their ideas in the past. Enactivism also has a long history, perhaps going back to the work of Chilean cognitive scientists Maturana and Varela. However, it is safe to argue that such literature is not to well studied by most AI practitioners.
David Marr described 3 layers of explanation for neuroscience. A very important one that is often overlooked is the ‘computational layer’. The computational layers explain why a brain does what it does. It is its intrinsic motivation. We can in fact plot the different AI approaches in a graph with a horizontal axis describing epistemology (i.e. how is knowledge arrived at) and a vertical axis describing ontology (i.e. models of the world).
The meta-physics underlying deep learning is in fact one that is enactivist in nature.
First I can’t discern if there’s a logical order in how these seven are presented. So allow me to re-order them in a manner that makes sense to me. That is from an Enactivist perspective.
Receptive — The ability to ‘appreciate’ existing conventions, and to receive new ones. Most Deep Learning approaches enable this. I favor another term for this, I would call this intuitive. In fact, it is counterintuitive to realize that our language ability is a consequence of System 1 (i.e. Intuition) thinking. I explain this further in my exploration of reasoning.
Graded — This is a very good insight that has not been explicitly expressed enough. An agent’s ability for symbolic thinking covers a wide spectrum. I argued once that being diagnosed for being infected with a virus isn’t a binary measures (i.e. positive versus negative). It is important to know how much viral load had been exposed to. In the same way, symbolic fluency has varying degrees of competence. It is not a single behavioral condition to that one has access like Chomsky’s Merge capability. Rather, it is constructively developed through education and experience. One’s ability for symbolic thinking is related to one’s reasoning ability.
Malleable — Hofstadter would characterize this as conceptual slippage. The Bongard problem illustrates the importance of conceptual slippage.
Separable — “A symbol user should be able to demonstrate partitioning of their understanding of a substrate and symbol”. I call this Symmetry Breaking.
Constructive — The ability to form new conventions by imposing new meaning on effectively arbitrary substrates. This implies a kind of compositionality that creates new meaning. This is absent in current AI.
Embedded — The ability to derive meaning from the context of which symbols are used. This relates to Wittgenstein’s exploration of language games.
Meaningful —
DeepMind’s paper ‘Symbolic Behaviour in Artificial Intelligence’ is certainly a good first step towards a framework of understanding of how to get to general intelligence.
I would like to end by an important quote from Alan Turing’s 1938 Ph.D. dissertation:
“We have gone to the opposite extreme and eliminated not intuition but ingenuity, and this in spite of the fact that our aim has been in much the same direction.”
What he describes as ingenuity is what we describe today as symbol manipulation. Computers didn’t exist in 1938, but Turing had the intuition that symbol manipulation could indeed be automated. However, he was not as certain as to how intuition could be automated.
Kudos to the DeepMind folks ( Adam Santoro, Andrew Lampinen, Kory Mathewson, Timothy Lillicrap, David Raposo ) for articulating a promising path towards general intelligence. It is good to know that I’m not the only ‘crazy’ person out there that recognizes this path. Hendricks-Jansen coined the neologism ‘interactive emergence’, I will follow in the same spirit and coin this as ‘symbolic emergence’.