Neuro-symbolic AI

Last updated

Neuro-symbolic AI is a type of artificial intelligence that integrates neural and symbolic AI architectures to address the weaknesses of each, providing a robust AI capable of reasoning, learning, and cognitive modeling. As argued by Leslie Valiant [1] and others, [2] [3] the effective construction of rich computational cognitive models demands the combination of symbolic reasoning and efficient machine learning. Gary Marcus, argued, "We cannot construct rich cognitive models in an adequate, automated way without the triumvirate of hybrid architecture, rich prior knowledge, and sophisticated techniques for reasoning." [4] Further, "To build a robust, knowledge-driven approach to AI we must have the machinery of symbol manipulation in our toolkit. Too much useful knowledge is abstract to proceed without tools that represent and manipulate abstraction, and to date, the only known machinery that can manipulate such abstract knowledge reliably is the apparatus of symbol manipulation." [5]

Contents

Henry Kautz, [6] Francesca Rossi, [7] and Bart Selman [8] also argued for a synthesis. Their arguments attempt to address the two kinds of thinking, as discussed in Daniel Kahneman's book Thinking Fast and Slow. It describes cognition as encompassing two components: System 1 is fast, reflexive, intuitive, and unconscious. System 2 is slower, step-by-step, and explicit. System 1 is used for pattern recognition. System 2 handles planning, deduction, and deliberative thinking. In this view, deep learning best handles the first kind of cognition while symbolic reasoning best handles the second kind. Both are needed for a robust, reliable AI that can learn, reason, and interact with humans to accept advice and answer questions. Such dual-process models with explicit references to the two contrasting systems have been worked on since the 1990s, both in AI and in Cognitive Science, by multiple researchers. [9]

Approaches

Approaches for integration are diverse. Henry Kautz's taxonomy of neuro-symbolic architectures, [10] along with some examples, follows:

These categories are not exhaustive, as they do not consider multi-agent systems. In 2005, Bader and Hitzler presented a more fine-grained categorization that considered, e.g., whether the use of symbols included logic and if it did, whether the logic was propositional or first-order logic. [14] The 2005 categorization and Kautz's taxonomy above are compared and contrasted in a 2021 article. [10] Recently, Sepp Hochreiter argued that Graph Neural Networks "...are the predominant models of neural-symbolic computing" [15] since "[t]hey describe the properties of molecules, simulate social networks, or predict future states in physical and engineering applications with particle-particle interactions." [16]

Artificial general intelligence

Gary Marcus argues that "...hybrid architectures that combine learning and symbol manipulation are necessary for robust intelligence, but not sufficient", [17] and that there are

...four cognitive prerequisites for building robust artificial intelligence:

This echoes earlier calls for hybrid models as early as the 1990s. [19] [20]

History

Garcez and Lamb described research in this area as ongoing at least since the 1990s. [21] [22] At that time, the terms symbolic and sub-symbolic AI were popular.

A series of workshops on neuro-symbolic AI has been held annually since 2005 Neuro-Symbolic Artificial Intelligence. [23] In the early 1990s, an initial set of workshops on this topic were organized. [19]

Research

Key research questions remain, [24] such as:

Implementations

Implementations of neuro-symbolic approaches include:

Citations

  1. Valiant 2008.
  2. Garcez et al. 2015.
  3. D'Avila Garcez, Artur S.; Lamb, Luis C.; Gabbay, Dov M. (2009). Neural-symbolic cognitive reasoning. Cognitive technologies. Springer. ISBN   978-3-540-73245-7.
  4. Marcus 2020, p. 44.
  5. Marcus & Davis 2019, p. 17.
  6. Kautz 2020.
  7. Rossi 2022.
  8. Selman 2022.
  9. Sun 1995.
  10. 1 2 Sarker, Md Kamruzzaman; Zhou, Lu; Eberhart, Aaron; Hitzler, Pascal (2021). "Neuro-symbolic artificial intelligence: Current trends". AI Communications. 34 (3): 197–209. doi:10.3233/AIC-210084. S2CID   239199144.
  11. Mao et al. 2019.
  12. Rocktäschel, Tim; Riedel, Sebastian (2016). "Learning Knowledge Base Inference with Neural Theorem Provers". Proceedings of the 5th Workshop on Automated Knowledge Base Construction. San Diego, CA: Association for Computational Linguistics. pp. 45–50. doi: 10.18653/v1/W16-1309 . Retrieved 2022-08-06.
  13. Serafini, Luciano; Garcez, Artur d'Avila (2016). "Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge". arXiv: 1606.04422 [cs.AI].
  14. Bader & Hitzler 2005.
  15. L.C. Lamb, A.S. d'Avila Garcez, M.Gori, M.O.R. Prates, P.H.C. Avelar, M.Y. Vardi (2020). "Graph Neural Networks Meet Neural-Symbolic Computing: A Survey and Perspective." CoRR abs/2003.00330 (2020)
  16. Hochreiter, Sepp (April 2022). "Toward a broad AI". Communications of the ACM. 65 (4): 56–57. doi:10.1145/3512715. ISSN   0001-0782.
  17. Marcus 2020, p. 50.
  18. Marcus 2020, p. 48.
  19. 1 2 Sun & Bookman 1994.
  20. Honavar 1995.
  21. Garcez & Lamb 2020, p. 2.
  22. Garcez et al. 2002.
  23. "Neuro-Symbolic Artificial Intelligence". people.cs.ksu.edu. Retrieved 2023-09-11.
  24. Sun 2001.
  25. Li, Ziyang; Huang, Jiani; Naik, Mayur (2023). "Scallop: A Language for Neurosymbolic Programming". arXiv: 2304.04812 [cs.PL].
  26. "Model Induction Method for Explainable AI". USPTO. 2021-05-06.

Related Research Articles

Artificial intelligence (AI), in its broadest sense, is intelligence exhibited by machines, particularly computer systems. It is a field of research in computer science that develops and studies methods and software that enable machines to perceive their environment and uses learning and intelligence to take actions that maximize their chances of achieving defined goals. Such machines may be called AIs.

<span class="mw-page-title-main">Cognitive science</span> Interdisciplinary scientific study of cognitive processes

Cognitive science is the interdisciplinary, scientific study of the mind and its processes. It examines the nature, the tasks, and the functions of cognition. Mental faculties of concern to cognitive scientists include language, perception, memory, attention, reasoning, and emotion; to understand these faculties, cognitive scientists borrow from fields such as linguistics, psychology, artificial intelligence, philosophy, neuroscience, and anthropology. The typical analysis of cognitive science spans many levels of organization, from learning and decision to logic and planning; from neural circuitry to modular brain organization. One of the fundamental concepts of cognitive science is that "thinking can best be understood in terms of representational structures in the mind and computational procedures that operate on those structures."

<span class="mw-page-title-main">Connectionism</span> Cognitive science approach

Connectionism is the name of an approach to the study of human mental processes and cognition that utilizes mathematical models known as connectionist networks or artificial neural networks. Connectionism has had many 'waves' since its beginnings.

<span class="mw-page-title-main">Symbolic artificial intelligence</span> Methods in artificial intelligence research

In artificial intelligence, symbolic artificial intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level symbolic (human-readable) representations of problems, logic and search. Symbolic AI used tools such as logic programming, production rules, semantic nets and frames, and it developed applications such as knowledge-based systems, symbolic mathematics, automated theorem provers, ontologies, the semantic web, and automated planning and scheduling systems. The Symbolic AI paradigm led to seminal ideas in search, symbolic programming languages, agents, multi-agent systems, the semantic web, and the strengths and limitations of formal knowledge and reasoning systems.

Neuromorphic computing is an approach to computing that is inspired by the structure and function of the human brain. A neuromorphic computer/chip is any device that uses physical artificial neurons to do computations. In recent times, the term neuromorphic has been used to describe analog, digital, mixed-mode analog/digital VLSI, and software systems that implement models of neural systems. The implementation of neuromorphic computing on the hardware level can be realized by oxide-based memristors, spintronic memories, threshold switches, transistors, among others. Training software-based neuromorphic systems of spiking neural networks can be achieved using error backpropagation, e.g., using Python based frameworks such as snnTorch, or using canonical learning rules from the biological learning literature, e.g., using BindsNet.

In artificial intelligence (AI), commonsense reasoning is a human-like ability to make presumptions about the type and essence of ordinary situations humans encounter every day. These assumptions include judgments about the nature of physical objects, taxonomic properties, and peoples' intentions. A device that exhibits commonsense reasoning might be capable of drawing conclusions that are similar to humans' folk psychology and naive physics.

The expression computational intelligence (CI) usually refers to the ability of a computer to learn a specific task from data or experimental observation. Even though it is commonly considered a synonym of soft computing, there is still no commonly accepted definition of computational intelligence.

A cognitive architecture refers to both a theory about the structure of the human mind and to a computational instantiation of such a theory used in the fields of artificial intelligence (AI) and computational cognitive science. The formalized models can be used to further refine a comprehensive theory of cognition and as a useful artificial intelligence program. Successful cognitive architectures include ACT-R and SOAR. The research on cognitive architectures as software instantiation of cognitive theories was initiated by Allen Newell in 1990.

<span class="mw-page-title-main">Sally–Anne test</span> Psychological test

The Sally–Anne test is a psychological test, used in developmental psychology to measure a person's social cognitive ability to attribute false beliefs to others. The flagship implementation of the Sally–Anne test was by Simon Baron-Cohen, Alan M. Leslie, and Uta Frith (1985); in 1988, Leslie and Frith repeated the experiment with human actors and found similar results.

Hybrid intelligent system denotes a software system which employs, in parallel, a combination of methods and techniques from artificial intelligence subfields, such as:

Dov M. Gabbay is an Israeli logician. He is Augustus De Morgan Professor Emeritus of Logic at the Group of Logic, Language and Computation, Department of Computer Science, King's College London.

The following outline is provided as an overview of and topical guide to artificial intelligence:

Ron Sun is a cognitive scientist who made significant contributions to computational psychology and other areas of cognitive science and artificial intelligence. He is currently professor of cognitive sciences at Rensselaer Polytechnic Institute, and formerly the James C. Dowell Professor of Engineering and Professor of Computer Science at University of Missouri. He received his Ph.D. in 1992 from Brandeis University.

Artur d'Avila Garcez is a researcher in the field of computational logic and neural computation, in particular hybrid systems with application in software verification and information extraction. His contributions include neural-symbolic learning systems and nonclassical models of computation combining robust learning and reasoning. He is a Professor of Computer Science at City, University London.

In the philosophy of artificial intelligence, GOFAI is classical symbolic AI, as opposed to other approaches, such as neural networks, situated robotics, narrow symbolic AI or neuro-symbolic AI. The term was coined by philosopher John Haugeland in his 1985 book Artificial Intelligence: The Very Idea.

This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.

<span class="mw-page-title-main">Differentiable neural computer</span> Artificial neural network architecture

In artificial intelligence, a differentiable neural computer (DNC) is a memory augmented neural network architecture (MANN), which is typically recurrent in its implementation. The model was published in 2016 by Alex Graves et al. of DeepMind.

Explainable AI (XAI), often overlapping with Interpretable AI, or Explainable Machine Learning (XML), either refers to an artificial intelligence (AI) system over which it is possible for humans to retain intellectual oversight, or refers to the methods to achieve this. The main focus is usually on the reasoning behind the decisions or predictions made by the AI which are made more understandable and transparent. XAI counters the "black box" tendency of machine learning, where even the AI's designers cannot explain why it arrived at a specific decision.

<span class="mw-page-title-main">Pascal Hitzler</span> German-American computer scientist

Pascal Hitzler is a German American computer scientist specializing in Semantic Web and Artificial Intelligence. He is endowed Lloyd T. Smith Creativity in Engineering Chair, one of the Directors of the Institute for Digital Agriculture and Advanced Analytics (ID3A) and Director of the Center for Artificial Intelligence and Data Science (CAIDS) at Kansas State University, and the founding Editor-in-Chief of the Semantic Web journal and the IOS Press book series Studies on the Semantic Web.

References

See also