Dual-process theories, cognitive architectures, and hybrid neural-symbolic models
Abstract
This position paper discusses relationships among hybrid neural-symbolic models, dual-process theories, and cognitive architectures. It provides some historical backgrounds and argues that dual-process (implicit versus explicit) theories have significant implications for developing neural-symbolic (neurosymbolic) models. Furthermore, computational cognitive architectures can help to disentangle issues concerning dual-process theories and thus help the development of neural-symbolic models (in this way as well as in other ways).
1.Introduction
In this relatively short position paper, I will briefly discuss hybrid neural-symbolic models, dual-process theories, and cognitive architectures, as well as mutually beneficial relationships among them and their relevance to each other. I will first provide some historical backgrounds. Based on that, I will argue that dual-process (implicit versus explicit) theories have significant implications for developing neural-symbolic (or neurosymbolic) models. Furthermore, through computational means, cognitive architectures can help to disentangle complex issues concerning dual-process theories and thus help with the further development of hybrid neural-symbolic models (in this as well as other ways). The present position paper inevitably reflects, of course, a personal perspective resulting from personal experiences, and is not meant to be comprehensive and complete (nor technically detailed).
Neural networks and deep learning are good at many things, but they have shortcomings and limitations, despite their popularity, as has long been known (e.g., difficulties with certain types of extrapolation, generalization, abstraction, explanation, and so on; see, e.g., very early arguments at the beginning of “new connectionism”, such as Fodor and Pylyshyn [11]; Marcus [27]; Pinker and Prince [33]).11 But, as many have noticed, neural networks and symbolic models have some complementary strengths. Thus, combining them in some way seems a good idea naturally, and the resulting models can be advantageous – For instance, they can be more expressive, more versatile, or more capable otherwise. Hence the idea of hybrid neural-symbolic (or neurosymbolic) models was in place, transcending many heated and conflicting claims of advantages and disadvantages of each paradigm (Sun [40,45]).
This idea immediately harkens back to the 1990s when hybrid models first emerged (see, e.g., Sun and Bookman [50]). At the present time (at the time when the present article is being written), this idea is clearly not something entirely new (or new at all). In other words, the old has become new again (in ways much like fashion, albeit with some necessary modifications and variations). Over time, such models have been known under various names: Besides being termed neural-symbolic or neurosymbolic models, they have also been variously known as connectionist-symbolic model, hybrid symbolic neural networks, or simply hybrid models or systems, at different times in their history (see, e.g., Sun [40,45]; Sun and Bookman [50]; Wermter and Sun [53]).
Despite the fact that this important (or even crucial) idea was very much ignored or downplayed early on by the mainstreams on both the neural and the symbolic side, by now, even some strong advocates for symbolic AI have suggested that hybrid neural-symbolic models should indeed be pursued (e.g., Kautz [21]). At the same time, some strong advocates for neural networks and deep learning have also started to see that (e.g., LeCun [24]). When people from the two opposing camps converge onto the same idea (along with people who have always been on that ground somewhere in the middle), the likelihood of this idea leading to major advances has just skyrocketed. (Some of them even invoked good old dual-process theories in support of their arguments; more on this later.) There have also been quite a number of recent events promoting this theme (which I shall not enumerate, for this would take far too long), including, in particular, the launch of the present journal.
However, there have been many possible ways of structuring such models; that is, there exist many different ways of combining symbolic processing and neural networks (Sun [40]). Many of them, however, appear ad hoc or task specific, or otherwise problematic. Therefore, a major question that immediately comes to mind in regard to such hybrid models is: how should we structure them after all? How do we structure them in a way that avoids being ad hoc? How do we make them more principled, that is, justifiable by some fundamental laws or principles? Obviously, it is preferable that they should have a more solid theoretical (as well as empirical) basis – that is, they should be justified by some first principles (or something very close to that).
2.Some early history (or what you might call pre-history)
Some early work on neural-symbolic models is worth mentioning here, due to their close relevance to what we have discussed thus far. Even during the very early days of such models (e.g., in the 1990s), there was some important work that had been done. One may choose to view such early work as the pre-history to today’s neural-symbolic models (especially when they involve complex deep neural networks that were not yet fully developed or popularized back then), or may just simply view it as the early history of this evolving field, thus conceiving a more continuous process of development and progression, In this regard, it should be noted that in the 1990s, there were some edited volumes (books) with contributions from many authors, which meant that they were likely representative of the state of the art of the work done during that period and contained many different models and ideas, some of which may still be relevant and useful today. Therefore, we shall have a look.
One of these books was Sun and Bookman [50] (which, by the way, happened to be the first such volume on the topic of neural-symbolic models). One of the contributions of this book (and some other related ones) is dealing with the analysis, classification, categorization, and cataloguing of neural-symbolic models (e.g., different ways of designing and constructing such models, different computational techniques used, and so on).
For example, some general categories discussed in Sun and Bookman [50] included: (1) designing specialized, structured, localist neural networks for symbolic processing (e.g., some forms of rule-based reasoning); (2) embedding symbolic processing within usual neural networks with distributed representations; (3) having separate neural and symbolic processing modules but connecting them; (4) using neural networks as elements (modules) in a structured, symbolic architecture (see Sun and Bookman [50] for details of these categories). On the other hand, from the perspective of a whole system, (1) a system could be a single-module system, in which (1.1) representations can be symbolic, localist, or distributed, and (1.2) mappings between symbols and representations could be direct, translational, or transformational (fully utilizing complex neural dynamics); or (2) it could be a multi-module system, which (2.1) could be either homogeneous or heterogeneous in terms of representation or structure and (2.2) could vary in terms of granularity of modules, and (2.3) the relationship between modules could vary, for example, either loosely or tightly coupled. There can also be finer distinctions and variations (Sun and Alexandre [49]). Symbolic reasoning, logic, and probabilistic reasoning could have various neural implementations or approximations. These ideas represented the state of the art in the 1990s, so, obviously, there are many newer methods and ideas by now (e.g., back then, there was no sophisticated deep learning or Big Data). However, note that these models and methods were designed to address the fundamental shortcomings of purely neural or purely symbolic models; these shortcomings were known even back then but remain a challenge today despite the progress that has been made. Many of the concepts described above (such as localist versus distributed representations, single- versus multi-module systems, loosely versus tightly coupled modules, and so on) and many common structures of neural-symbolic systems identified back then (such as direct mapping of symbolic models to neural networks, embedding symbolic processing within neural networks, multiple neural and/or symbolic modules with different control and coordination methods, and so on) remain viable in describing today’s neural-symbolic models.
However, one question that one faced then, and is still facing today, is: Given these options, that is, given these many different ways of combining symbolic processing and neural networks, structurally, mechanistically, and process-wise, how can we produce better hybrid models going forward? That is, how should we best structure and implement them, choosing among all these options? In particular, how do we structure them in a principled and justifiable way, rather than in an ad hoc or task-specific way?
I argued back then and am still arguing today (e.g., in some of my earlier or more recent books, such as Sun [38,41,47]) that a better approach is structuring them in a cognitively (i.e., psychologically) motivated and justified way – that is, based on human mental architecture (which is, in turn, inferred based on experimental psychology, neuroscience, and so on; see Newell [31]; Sun [42]). In particular, within the human mental architecture, we need to take into account dual processes (e.g., as has been variously termed as implicit versus explicit, unconscious versus conscious, intuition versus reason, System 1 versus System 2, and so on, albeit sometimes with somewhat different connotations). Incidentally, dual-process (or two-system) theories have become quite popular lately (see, e.g., Kahneman [19]), although they are not new ideas either (see, e.g., Reber [34]; Sun [42]; and others). Dual-process theories can provide both philosophical and psychological justifications for (at least some of) hybrid neural-symbolic models. In fact, they constitute both theoretical and empirical grounding for hybrid neural-symbolic models.
3.Dual-process theories
So, at this juncture, we should examine (however briefly) dual-process theories that have been proposed in the literature – There have been quite a few of them out there, and their historical roots go even deeper. It is worthwhile to look into them here for the sake of better understanding of neural-symbolic models.
Theoretical or philosophical proposals concerning two types of mental processes have long existed, dating back to even before the inception of cognitive science and artificial intelligence. For instance, Martin Heidegger’s distinction of the pre-ontological versus the ontological is an early, very abstract version of such a duality (Heidegger [15]). William James’s [18] distinction between “empirical thinking” and “true reasoning” is even more evidently relevant. Going back even earlier, Immanuel Kant’s [20] dictum was: “Thoughts without intuition are empty, intuition without concepts is blind”. “Intuition and concepts constitute... the elements of all our knowledge, so that neither concepts without an intuition in some way corresponding to them, nor intuition without concepts, can yield knowledge” (Kant [20]).
In contemporary experimental psychology and neuroscience, there have been empirically based proposals in this regard. The distinction of implicit and explicit processes has been empirically demonstrated in the implicit memory literature (e.g., Schacter [35]). The distinction of implicit and explicit processes has also been empirically demonstrated in the implicit learning literature (e.g., Reber [34]). In social psychology, there have been dual-process models that are roughly based on the co-existence of implicit and explicit processes (see Chaiken and Trope [3] for some examples). Similar distinctions have been proposed by other researchers, based on similar or different empirical or theoretical considerations (e.g., Erickson and Kruschke [8]; Grossberg [14]; McClelland, McNaughton, and O’Reilly [28]; Milner and Goodale [30]; Sun [38,39]; etc.).
Here is my version of a dual-process theory from a long time ago. In my 1994 book, I made the following distinction:
“
That is, the two “levels” (the two systems or modules, or two types of processes) encode somewhat similar content, but they encode their content in different ways: Symbolic versus subsymbolic representations are used, respectively. Therefore, they utilize different mechanisms. They can thus have qualitatively different “flavors”, including different degrees of accessibility (leading to different degrees of conscious awareness). One may be termed the implicit level, and the other the explicit level (or, alternatively, they may be termed System 1 and System 2, respectively, or whatever other terms one may wish to use as mentioned earlier).
Furthermore, according to this account, a highly plausible reason for having the two different levels is that these different levels (with different representations and different mechanisms) can potentially work together synergistically, complementing and supplementing each other (as argued, in detail, in Sun [38,39,41], and Sun et al. [51]; see also Booch et al. [2] and Kautz [21] for more recent views). This is probably, at least in part, the reason why there are these two levels – This likely contributes to why nature has chosen this design (in an evolutionary sense, of course). I referred to the above two points as the dual representation hypothesis and the synergy hypothesis, respectively (see Sun [41,46]).
A number of other dual-process theories exist, although some of them can be problematic sometimes. For instance, a comparatively more recent view (compared with those precedent views mentioned earlier) was proposed by Kahneman (e.g., [19]). The main ideas of his were as follows: There are two styles of processing: intuition and reasoning. Intuition (or System 1) is based on associative processes, fast and automatic, involving strong emotional bonds, based on formed habits, and difficult to change or manipulate. Reasoning (or System 2) is slower, more volatile, and subject to conscious judgments and attitudes.
At roughly around the same time, Evans [9] espoused essentially the same view. According to Evans, System 1 is “rapid, parallel and automatic in nature: only their final product is posted in consciousness”. He also noted the “domain-specific nature” of System 1 learning. System 2, on the other hand, is “slow and sequential in nature and makes use of the central working memory system”; it “permits abstract hypothetical thinking that cannot be achieved by System 1”. Further, there is the “default-intervention” relationship between the two systems: System 1 forms the default response unless there is active intervention from System 2
However, it should be worth noting that some of these claims above, unfortunately, may be somewhat simplistic. For one thing, intuition can be slow sometimes (see, e.g., Helie and Sun [16]). For another, intuition can be subject to conscious control and manipulation; that is, it may not be entirely “automatic” (e.g., Curran and Keele [5]; Stadler [37]). Furthermore, intuition can be subject to conscious “judgment” (e.g., Libet [25]). Moreover, explicit thinking may be engaged right from the outset and can be constantly present, not necessarily just an occasional intervention, and there can be complex interactions between the two systems (Helie and Sun [16]). The reader is referred to Sun [44,46] for detailed discussions of these points. For various alternative dual-process views, see also Evans and Frankish [10] and Macchi et al. [26].
To come up with a more nuanced and more precise characterization of the two systems, it is important that we ask some pertinent and deeper questions, taking into account architectural and mechanistic nuances of the mind. For instance, for either type of process, there can be the following important questions (that are mechanistic and/or process-oriented):
How deep is the processing (in terms of precision, certainty, and so on)?
How broad is the processing (e.g., how much information is involved)?
How degraded (incomplete, inconsistent, or uncertain) is the information available?
How typical or atypical is the situation to be addressed?
Is the process involved procedural or declarative?
How many processing cycles are needed considering the factors above?
There are many other similar or related questions, including those related to further divisions of modules within each level and across levels, such as procedural versus declarative modules (see, e.g., Anderson and Lebiere [1]) and their relations to the implicit-explicit distinction. (Note that the relationship between System 1 versus 2 and procedural versus declarative processes is a topic that is worth an article-length or even a book-length treatment by itself. It is certainly not simply the case that System 2 involves only declarative processes with complex knowledge representation, while System 1 involves only procedural processes with simple learning, as some researchers in the field seem to believe. See Sun [43] for details.)
These questions need to be addressed, to ensure cognitive-psychological realism of theories or models. To put it in a different way, the complexity of the matter should be adequately understood and should be properly taken into account, especially in any mechanistic models or theories that purport to value cognitive-psychological realism (Sun [46]).22 Hybrid neural-symbolic models belong (or, at least, should belong) to this category of models.
4.Computational cognitive architectures
To better understand relevant issues and to better answer these questions raised earlier, one would certainly like to have an overall theoretical framework, as well as detailed computational (mechanistic) models that can substantiate proposed theories. That is where computational cognitive architectures (Newell [31]) come into play – because they are both types of entities (as mentioned above) at the same time and thus they can go a long way in helping to clarify these issues. Note that the notion of computational cognitive architecture, as developed within cognitive science, refers to a broadly scoped, domain-generic computational cognitive (i.e., psychological) model, capturing the essential structures, mechanisms, representations, and processes of the mind, to be used for broadly scoped, multi-domain, multi-scale analysis of behavior (Sun [47]). Needless to say, they are meant to be cognitively-psychologically realistic.
Computational cognitive architectures provide an overall framework that facilitates more detailed explorations of the mind (e.g., through providing conceptual tools and constraints; Laird et al. [22]; Taatgen and Anderson [52]). This general framework can then lead to specific theories about components or functionalities of the mind. We accomplish the above through specifying computational details of cognitive (psychological) mechanisms and processes; in other words, we embody our theories of the mind and its various functionalities in computer programs. In so doing, we generate runnable computer programs, which enables the simulation of empirical human data; simulation results can then be compared to human data, in a qualitative or quantitative way, to validate the theories or to develop better ones. Through this iterative and detailed process of comparing human data and simulations, we can achieve better understanding of the inner workings of the human mind. This is the essence of computational psychology (Sun [42]). With cognitive architectures, we focus first on a general theory (a general framework), then move on to more specific theories (Newell [31]).
As an example, the Clarion cognitive architecture is a dual-process, dual-representation cognitive architecture, incorporating both implicit and explicit processes. It uses dual representations, that is, both symbolic and neural representations, for computationally capturing and differentiating explicit and implicit processes (see Sun [41,47] regarding Clarion; see also Sun [38] for ideas leading up to it). Clarion has been aimed to be an integrative and comprehensive computational framework. For instance, Clarion encompasses many different mechanisms, covers many different psychological functionalities and domains, and at the same time develops a more nuanced view of the two systems (the two processes/levels). Understanding intelligence and cognition (both natural and artificial kinds), especially issues related to dual processes, requires no less than such a comprehensive model. But, in turn, dual processes are crucial for this and many other cognitively-psychologically realistic models, serving as their theoretical inspiration and both theoretical and empirical justifications.
One can view modeling and simulation based on cognitive architectures as a way of theoretical interpretation (albeit in a detailed and precise, namely mechanistic, fashion). In this way, Clarion has indeed helped to clarify many issues related to dual processes mentioned earlier. For instance, some claimed that implicit processes (“intuition”) are always faster than explicit processes (“reasoning”). In terms of procedural processes, this is generally true, using theoretical interpretation through simulation (e.g., Sun et al. [51]). In terms of declarative processes, however, this is generally not true (e.g., Helie and Sun [16]). See Sun ([44,46]) for analysis of these and other issues related to dual-process theories and beyond.
The established methodology of developing computational cognitive architectures on the basis of empirical data and findings from psychology, neuroscience, and other disciplines, constitutes, at the same time, a principled (and thus arguably preferred) way of developing neural-symbolic systems in general, regardless of whether these systems are for theoretical purposes or for practical applications (cf. Pew and Mavor [32]; Sun [42]). In relation to practical applications, the advantages of this approach include, for example, that a system resulting from this methodology will likely be more similar to humans in some fundamental way, that it is possible that such a system is more easily understood by humans (and vice versa), and that it may be easier for such a system to communicate with humans, not just through language, but also through other explicit or implicit means of communication (Sun [48]).
5.Some further remarks
Without going further into details of how dual processes can be (and indeed have been) further developed conceptually and mechanistically within Clarion and in turn help the development of neural-symbolic models (which is due to the space limit, for one thing, but also due to the nature of a position paper such as this), I would like to simply highlight a few key points:
Dual-process theories have significant implications for developing computational cognitive architectures and cognitively-psychologically realistic neural-symbolic models (Sun [38,40,41]).
If cognitive-psychological realism is what one wants to achieve in developing theories or models, dual-process theories must be taken into consideration, not just as an option, not merely for expediency, but as a rather fundamental requirement (Sun [46]). (Of course, it is obvious that cognitive-psychological realism may not be for everyone, especially those who are only concerned with AI applications in the short run. But, even in those cases, there have been some calls for dual-process theories to be applied, although mostly for the sake of practical expediency.)
Dual-process theories can serve as the inspiration (and both theoretical and empirical justifications) for computational cognitive architectures and neural-symbolic models in general (especially cognitively-psychologically motivated ones).
Recent developments of deep learning (but also see, e.g., Schmidhuber [36] for its long history) have been exciting, especially the successes that convolutional neural networks, transformers, large language models based on transformers, computer vision based on deep neural networks, and other similar fronts have been enjoying. Many ideas that have been developing over the past three to four decades came together in a metaphoric big bang. Of course, such developments have also introduced many new issues and many new possibilities in relation to neural-symbolic models (Garcez et al. [13]; Garcez and Jiménez-Ruiz [12]; Hitzler et al. [17]), including, for example, probabilistic approaches towards integrating neural and symbolic processes, reasoning on the basis of large language models, and so on. However, there are reasons to believe that these fundamental points that were made earlier (such as these listed above) still stand (although adjustments and adaptations might be needed along the way naturally). It should be pointed out that the various recent calls (in some presidential speeches no less) for attention to the distinction between System 1 and System 2 in developing artificial intelligence systems involving deep learning, although timely, are nothing new either – Systems and models with the same or similar distinctions have been developed at one point or another, or (in some cases) have been in continuous development, non-stop, since the 1990s. Such ideas have stood the test of time.
Going back to the points and categories mentioned earlier in our brief review of the early history of neural-symbolic models, dual-process theories point to multi-module, multi-representational, heterogeneous neural-symbolic systems as the basis of the human mental architecture (as discussed in the early years in Sun and Bookman [50]; Sun and Alexandre [49]; Wermter and Sun [53]). Further empirical and/or theoretical support pointing in that direction has been mounting over the past several decades since then.
It is worth repeating that, despite the fact that dual-process theories are quite popular right now (as clearly indicated by the recent surge of interests in System 1 and System 2), some important issues involved in dual-process theories (e.g., relative speed, intention, control, and interaction) are more complex than often assumed. In this regard, it is essential to gain a more fine-grained understanding, and further work is needed to disentangle these issues (Sun [46]). These issues, however, are crucial in further developing cognitive architectures (either enhancing existing ones or starting anew, but beyond just a standard model; cf. Laird et al. [22]), and in turn, cognitive architectures can help in disentangling these and other theoretically important issues, by providing an overarching framework as well as by enabling computational simulation. Together, they may lead to better neural-symbolic models in the future; in other words, they may guide the development of future neural-symbol systems that are principled and cognitively-psychologically realistic. Nevertheless, there is, of course, still a long way to go, and new breakthroughs, incremental developments, enhancements, and refinements are very much needed (for some of the more recent developments, see, e.g., Chaudhuri et al. [4]; Dong [6]; Dong et al. [7]; Lamb et al. [23]; Mekik [29]; amongst many others).
Notes
1 To that list of shortcomings, one may also add difficulties with catastrophic forgetting, transfer of learning, distributional changes, amount of data required, computational cost, and so on, as many researchers discovered over time.
2 Few things in science are uncontroversial. There is a plethora of evidence that the two separate levels exist, although the point of two separate levels itself is not yet universally accepted (i.e., there are still people, existing today as well as in the 1990s, who dispute such a point – therefore it is a set of “theories”). Furthermore, the fine-grained details (e.g., representations and mechanisms) of the two levels and how they operate exactly process-wise are still under investigation (despite the fact that there have been some cogent theories, as mentioned earlier). There is also a tension between ontology (the dichotomy) and epistemology (finer details) in this regard.
Acknowledgement
The present article was written while the author was supported (in part) by ARI grant W911NF-17-1-0236 and IARPA HIATUS contract 2022-22072200002. The views and conclusions contained herein are those of the author’s and should not be interpreted as necessarily representing the official policies and positions, either expressed or implied, of those agencies. Thanks are due to the three anonymous reviewers for their useful suggestions.
References
[1] | J.R. Anderson and C. Lebiere, The Atomic Components of Thought, Lawrence Erlbaum Associates, Mahwah, NJ, (1998) . |
[2] | G. Booch, F. Fabiano, L. Horesh, K. Kate, J. Lenchner, N. Linck, A. Loreggia, K. Murgesan, N. Mattei, F. Rossi and B. Srivastava, Thinking fast and slow in AI, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35: , (2021) , pp. 15042–15046. |
[3] | S. Chaiken and Y. Trope (eds), Dual Process Theories in Social Psychology, Guilford Press, New York, (1999) . |
[4] | S. Chaudhuri, K. Ellis, O. Polozov, R. Singh, A. Solar-Lezama and Y. Yue, Neurosymbolic programming, Foundations and Trends in Programming Languages 7: (3) ((2021) ), 158–243. doi:10.1561/2500000049. |
[5] | T. Curran and S. Keele, Attention and structure in sequence learning, Journal of Experimental Psychology: Learning, Memory, and Cognition 19: ((1993) ), 189–202. |
[6] | T. Dong, A Geometric Approach to the Unification of Symbolic Structures and Neural Networks, Springer, Berlin, (2021) . |
[7] | T. Dong, A. Rettinger, J. Tang, B. Tversky and F. van Harmelen (eds), Structure and Learning. Dagstuhl Reports, Vol. 11: , Schloss Dagstuhl, Leibniz-Zentrum fur Informatik, Dagstuhl, Germany, (2022) , pp. 11–34. doi:10.4230/DagRep.11.8.11. |
[8] | M. Erickson and J. Kruschke, Rules and exemplars in category learning, Journal of Experimental Psychology: General 127: ((1998) ), 107–140. doi:10.1037/0096-3445.127.2.107. |
[9] | J. Evans, In two minds: Dual-process accounts of reasoning, Trends in Cognitive Sciences 7: (10) ((2003) ), 454–459. doi:10.1016/j.tics.2003.08.012. |
[10] | J. Evans and K. Frankish (eds), Two Minds: Dual Processes and Beyond, Oxford University Press, Oxford, UK, (2009) . doi:10.1093/acprof:oso/9780199230167.001.0001. |
[11] | J.A. Fodor and Z.W. Pylyshyn, Connectionism and cognitive architecture: A critical analysis, Cognition 28: (1–2) ((1988) ), 3–71. doi:10.1016/0010-0277(88)90031-5. |
[12] | A. Garcez and E. Jiménez-Ruiz, Neural-Symbolic Learning and Reasoning: Proceedings of the 16th International Workshop on Neural-Symbolic Learning and Reasoning, Windsor Great Park, United Kingdom, September 28–30, (2022) . |
[13] | A. Garcez, T. Besold, L. De Raedt, P. Földiák, P. Hitzler, T. Icard, K.-U. Kühnberger, L. Lamb, R. Miikkulainen and D. Silver, Neural-Symbolic Learning and Reasoning: Contributions and Challenges. AAAI Spring Symposium on Knowledge Representation and Reasoning: Integrating Symbolic and Neural Approaches, AAAI Press, Palo Alto, CA, (2015) , Technical Report SS-15-03. |
[14] | S. Grossberg, Nonlinear neural networks: Principles, mechanisms, and architectures, Neural Networks 1: ((1988) ), 17–61. doi:10.1016/0893-6080(88)90021-4. |
[15] | M. Heidegger, Being and Time, (1927) , English translation published by Harper and Row, New York. 1962. |
[16] | S. Helie and R. Sun, Incubation, insight, and creative problem solving: A unified theory and a connectionist model, Psychological Review 117: (3) ((2010) ), 994–1024. doi:10.1037/a0019532. |
[17] | P. Hitzler et al., Neuro-symbolic approaches in artificial intelligence, National Science Review ((2022) ). |
[18] | W. James, The Principles of Psychology, Dover, New York, (1890) . |
[19] | D. Kahneman, Thinking, Fast and Slow, Farrar, Straus and Giroux, New York, (2011) . |
[20] | I. Kant, Kritik der reinen Vernunft (Critique of Pure Reason), (1781) . |
[21] | H. Kautz, The Third AI Summer, Robert S. Engelmore Memorial Lecture. Thirty-Fourth AAAI Conference on Artificial Intelligence, New York, NY, (2020) , https://www.cs.rochester.edu/u/kautz/talks/Kautz%20Engelmore%20Lecture.pdf. |
[22] | J.E. Laird, C. Lebiere and P.S. Rosenbloom, A standard model of the mind: Toward a common computational framework across artificial intelligence, cognitive science, neuroscience, and robotics, AI Magazine 38: (4) ((2017) ), 13–26. doi:10.1609/aimag.v38i4.2744. |
[23] | L. Lamb, A. Garcez, M. Gori, M. Prates, P. Avelar and M. Vardi, Graph neural networks meet neural-symbolic computing: A survey and perspective, in: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, (2020) , pp. 4810–4817. doi:10.24963/ijcai.2020/670. |
[24] | Y. LeCun, A path towards autonomous machine intelligence, Open Review ((2022) ). |
[25] | B. Libet, Unconscious cerebral initiative and the role of conscious will in voluntary action, Behavioral and Brain Sciences 8: ((1985) ), 529–566. doi:10.1017/S0140525X00044903. |
[26] | L. Macchi, M. Bagassi and R. Viale (eds), Cognitive Unconscious and Human Rationality, MIT Press, Cambridge, MA, (2016) . |
[27] | G.F. Marcus, The Algebraic Mind: Integrating Connectionism and Cognitive Science, MIT Press, Cambridge, MA, (2001) . |
[28] | J. McClelland, B. McNaughton and R. O’Reilly, Why there are complementary learning systems in the hippocampus and neocortex: Insights from the successes and failures of connectionist models of learning and memory, Psychological Review 102: (3) ((1995) ), 419–457. doi:10.1037/0033-295X.102.3.419. |
[29] | C. Mekik, Raven’s Matrices in The Clarion Cognitive Architecture. Ph.D. Dissertation, Cognitive Science Department, Rensselaer Polytechnic Institute, (2022) . |
[30] | D. Milner and M. Goodale, The Visual Brain in Action, Oxford University Press, Oxford, UK, (1995) . |
[31] | A. Newell, Unified Theories of Cognition, Harvard University Press, Cambridge, MA, (1990) . |
[32] | R.W. Pew and A.S. Mavor (eds), Modeling Human and Organizational Behavior: Application to Military Simulations, National Academy Press, Washington, DC, (1998) . |
[33] | S. Pinker and A. Prince, On language and connectionism: Analysis of a parallel distributed processing model of language acquisition, Cognition 28: (1–2) ((1988) ), 73–193. doi:10.1016/0010-0277(88)90032-7. |
[34] | A. Reber, Implicit learning and tacit knowledge, Journal of Experimental Psychology: General. 118: (3) ((1989) ), 219–235. doi:10.1037/0096-3445.118.3.219. |
[35] | D. Schacter, Implicit memory: History and current status, Journal of Experimental Psychology: Learning, Memory, and Cognition 13: ((1987) ), 501–518. |
[36] | J. Schmidhuber, Deep learning in neural networks: An overview, Neural Networks 61: ((2015) ), 85–117. doi:10.1016/j.neunet.2014.09.003. |
[37] | M.A. Stadler, Role of attention in implicit learning, Journal of Experimental Psychology: Learning, Memory, and Cognition 21: ((1995) ), 674–685. |
[38] | R. Sun, Integrating Rules and Connectionism for Robust Commonsense Reasoning, John Wiley and Sons, New York, (1994) . |
[39] | R. Sun, Robust reasoning: Integrating rule-based and similarity-based reasoning, Artificial Intelligence (AIJ) 75: (2) ((1995) ), 241–296. doi:10.1016/0004-3702(94)00028-Y. |
[40] | R. Sun, Hybrid connectionist symbolic systems, in: Handbook of Brain Theories and Neural Networks, M. Arbib, ed., 2nd edn, MIT Press, Cambridge, MA, (2002) , pp. 543–547. |
[41] | R. Sun, Duality of the Mind, Lawrence Erlbaum Associates, Mahwah, NJ, (2002) . |
[42] | R. Sun (ed.), The Cambridge Handbook of Computational Psychology, Cambridge University Press, New York, (2008) . |
[43] | R. Sun, Memory systems within a cognitive architecture, New Ideas in Psychology 30: ((2012) ), 227–240. doi:10.1016/j.newideapsych.2011.11.003. |
[44] | R. Sun, Interpreting psychological notions: A dual-process computational theory, Journal of Applied Research in Memory and Cognition 4: (3) ((2015) ), 191–196. doi:10.1016/j.jarmac.2014.09.001. |
[45] | R. Sun, Artificial intelligence: Connectionist and symbolic approaches, in: International Encyclopedia of the Social and Behavioral Sciences, J.D. Wright, ed., Vol. 2: , 2nd edn, Elsevier, Oxford, UK, (2015) , pp. 35–40. doi:10.1016/B978-0-08-097086-8.43005-9. |
[46] | R. Sun, Implicit and explicit processes: Their relation, interaction, and competition, in: Cognitive Unconscious and Human Rationality, L. Macchi, M. Bagassi and R. Viale, eds, MIT Press, Cambridge, MA, (2016) . |
[47] | R. Sun, Anatomy of the Mind: Exploring Psychological Mechanisms and Processes with the Clarion Cognitive Architecture, Oxford University Press, New York, (2016) . |
[48] | R. Sun, Full human-machine symbiosis and truly intelligent cognitive systems, AI and Society: Journal of Knowledge, Culture and Communication 35: (1) ((2020) ), 17–28. doi:10.1007/s00146-017-0775-7. |
[49] | R. Sun and F. Alexandre (eds), Connectionist Symbolic Integration, Lawrence Erlbaum Associates, Hillsdale, NJ, (1997) . |
[50] | R. Sun and L. Bookman (eds), Computational Architectures Integrating Neural and Symbolic Processes, Kluwer Academic Publishers, Amsterdam, The Netherland, (1994) . |
[51] | R. Sun, P. Slusarz and C. Terry, The interaction of the explicit and the implicit in skill learning: A dual-process approach, Psychological Review 112: (1) ((2005) ), 159–192. doi:10.1037/0033-295X.112.1.159. |
[52] | N. Taatgen and J. Anderson, Chapter 8: Constraints in cognitive architectures, in: The Cambridge Handbook of Computational Cognitive Sciences, Cambridge University Press, Cambridge, UK, (2023) . |
[53] | S. Wermter and R. Sun (eds), Hybrid Neural Systems, Springer-Verlag, Berlin, (2000) . |