GOFAI
In the philosophy of artificial intelligence, GOFAI ("Good old fashioned artificial intelligence") is classical symbolic AI, as opposed to other approaches, such as neural networks, situated robotics, narrow symbolic AI or neuro-symbolic AI.[1][2] The term was coined by philosopher John Haugeland in his 1985 book Artificial Intelligence: The Very Idea.[3]
Haugeland coined the term to address two questions:
- Can GOFAI produce human level artificial intelligence in a machine?
- Is GOFAI the primary method that brains use to display intelligence?
AI founder Herbert A. Simon speculated in 1963 that the answers to both these questions was "yes". His evidence was the performance of programs he had co-written, such as Logic Theorist and the General Problem Solver, and his psychological research on human problem solving.[4]
AI research in the 1950s and 60s had an enormous influence on intellectual history: it inspired the cognitive revolution, led to the founding of the academic field of cognitive science, and was the essential example in the philosophical theories of computationalism, functionalism and cognitivism in ethics and the psychological theories of cognitivism and cognitive psychology. The specific aspect of AI research that led to this revolution was what Haugeland called "GOFAI".
Western rationalism
[edit]Haugeland places GOFAI within the rationalist tradition in western philosophy, which holds that abstract reason is the "highest" faculty, that it is what separates man from the animals, and that it is the most essential part of our intelligence. This assumption is present in Plato and Aristotle, in Shakespeare, Hobbes, Hume and Locke, it was central to the Enlightenment, to the logical positivists of the 1930s, and to the computationalists and cognitivists of the 1960s. As Shakespeare wrote:
What a piece of work is a man, How noble in reason, how infinite in faculty ... In apprehension how like a god, The beauty of the world, The paragon of animals.[5]
Symbolic AI in the 1960s was able to successfully simulate the process of high-level reasoning, including logical deduction, algebra, geometry, spatial reasoning and means-ends analysis, all of them in precise English sentences, just like the ones humans used when they reasoned. Many observers, including philosophers, psychologists and the AI researchers themselves became convinced that they had captured the essential features of intelligence. This was not just hubris or speculation -- this was entailed by rationalism. If it was not true, then it brings into question a large part of the entire Western philosophical tradition.
Continental philosophy, which included Nietzsche, Husserl, Heidegger and others, rejected rationalism and argued that our high-level reasoning was limited, prone to error, and that most of our abilities come from our intuitions, our culture, and from our instinctive feel for the situation. Philosophers who were familiar with this tradition were the first to criticize GOFAI and the assertion that it was sufficient for intelligence, such as Hubert Dreyfus and Haugeland.
Haugeland's GOFAI
[edit]Critics and supporters of Haugeland's position, from philosophy, psychology, or AI research have found it difficult to define "GOFAI" precisely, and thus the literature contains a variety of interpretations. Drew McDermott, for example, finds Haugeland's description of GOFAI "incoherent" and argues that GOFAI is a "myth".[6]
Haugeland coined the term GOFAI in order to examine the philosophical implications of “the claims essential to all GOFAI theories”,[3] which he listed as:
1. our ability to deal with things intelligently is due to our capacity to think about them reasonably (including sub-conscious thinking); and
2. our capacity to think about things reasonably amounts to a faculty for internal “automatic” symbol manipulation
— Haugeland (1985, p. 113)
This is very similar to the sufficient side of the physical symbol systems hypothesis proposed by Herbert A. Simon and Allen Newell in 1963:
"A physical symbol system has the necessary and sufficient means for general intelligent action."
— Newell & Simon (1976, p. 116)
It is also similar to Hubert Dreyfus' "psychological assumption":
"The mind can be viewed as a device operating on bits of information according to formal rules. "
— Dreyfus (1979, p. 157)
Haugeland's description of GOFAI refers to symbol manipulation governed by a set of instructions for manipulating the symbols. The "symbols" he refers to are discrete physical things that are assigned a definite semantics -- like <cat> and <mat>. They do not refer to signals, or unidentified numbers, or matrixes of unidentified numbers, or the zeros and ones of digital machinery.[7][8] Thus, Haugeland's GOFAI does not include "good old fashioned" techniques such as cybernetics, perceptrons, dynamic programming or control theory or modern techniques such as neural networks or support vector machines.
These questions ask if GOFAI is sufficient for general intelligence -- they ask if there is nothing else required to create fully intelligent machines. Thus GOFAI, for Haugeland, does not include systems that combine symbolic AI with other techniques, such as neuro-symbolic AI, and also does not include narrow symbolic AI systems that are designed only to solve a specific problem and are not expected to exhibit general intelligence.
Replies
[edit]This section needs expansion. You can help by adding to it. (July 2023) |
Replies from AI Scientists
[edit]Russell and Norvig wrote, in reference to Dreyfus and Haugeland:
The technology they criticized came to be called Good Old-Fashioned AI (GOFAI). GOFAI corresponds to the simplest logical agent design ... and we saw ... that it is indeed difficult to capture every contingency of appropriate behavior in a set of necessary and sufficient logical rules; we called that the qualification problem.[9]
Later symbolic AI work after the 1980's incorporated more robust approaches to open-ended domains such as probabilistic reasoning, non-monotonic reasoning, and machine learning.
Currently, most AI researchers [citation needed] believe deep learning, and more likely, a synthesis of neural and symbolic approaches (neuro-symbolic AI), will be required for general intelligence.
Citations
[edit]- ^ Boden 2014.
- ^ Segerberg, Meyer & Kracht 2020.
- ^ a b Haugeland 1985, p. 113.
- ^ Newell & Simon 1963.
- ^ Shakespeare, William. The Globe illustrated Shakespeare. The complete works, annotated, Deluxe Edition, (1986). Hamlet, Act II, scene 2, page 1879. Greenwich House, Inc. a division of Arlington House, Inc. distributed by Crown Publishers, Inc., 225 Park Avenue South, New York, NY 10003, USA.
- ^ Drew McDermott (2015), GOFAI Considered Harmful (And Mythical), S2CID 57866856
- ^ Touretzky & Pomerleau 1994.
- ^ Nilsson 2007, p. 10.
- ^ Russell & Norvig 2021, p. 982.
References
[edit]- Haugeland, John (1985), Artificial Intelligence: The Very Idea, Cambridge, Mass: MIT Press, ISBN 0-262-08153-9
- Boden, Margaret (2014), "GOFAI", in Keith Frankish; William M. Ramsay (eds.), The Cambridge Handbook of Artificial Intelligence, Cambridge University Press, pp. 89–107, ISBN 9781139046855,
Good Old-Fashioned AI – GOFAI, for short – is a label used to denote classical, symbolic, AI. The term "AI" is sometimes used to mean only GOFAI, but that is a mistake. AI also includes other approaches, such as connectionism (of which there are several varieties: see Chapter 5), evolutionary programming, and situated and evolutionary robotics.
- Segerberg, Krister; Meyer, John-Jules; Kracht, Marcus (Summer 2020), "The Logic of Action", in Zalta, Edward N. (ed.), The Stanford Encyclopedia of Philosophy,
[T]here is a tradition within AI to try and construct these systems based on symbolic representations of all relevant factors involved. This tradition is called symbolic AI or 'good old-fashioned' AI (GOFAI).
- Newell, Allen; Simon, H. A. (1963), "GPS: A Program that Simulates Human Thought", in Feigenbaum, E.A.; Feldman, J. (eds.), Computers and Thought, New York: McGraw-Hill
- Touretzky, David S.; Pomerleau, Dean A. (1994), "Reconstructing Physical Symbol Systems", Cognitive Science, 18 (2): 345–353, doi:10.1207/s15516709cog1802_5
- Nilsson, Nils (2007), Lungarella, M. (ed.), "The Physical Symbol System Hypothesis: Status and Prospects" (PDF), 50 Years of AI, Festschrift, LNAI 4850, Springer, pp. 9–17
- Russell, Stuart J.; Norvig, Peter. (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 9780134610993. LCCN 20190474.
- Dreyfus, Hubert (1979), What Computers Still Can't Do, New York: MIT Press.
- Newell, Allen; Simon, H. A. (1976), "Computer Science as Empirical Inquiry: Symbols and Search", Communications of the ACM, 19 (3): 113–126, doi:10.1145/360018.360022