Neats and scruffies
This article may require copy editing for grammar, style, cohesion, tone, or spelling. (August 2021) |
Neat and scruffy are styles of artificial intelligence (AI) research. The terms are also used as descriptions of developers or advocates whose work falls into one or the other of these two styles. Neats consider that solutions should be elegant, clear, and provably correct. Scruffies believe that intelligence is too complicated—or computationally intractable—to be solved with the sorts of homogeneous systems that neats usually mandate.
History
The distinction between neat and scruffy originated in the mid-1970s, by Roger Schank. Schank used the terms to characterize the difference between his work on natural language processing (which represented commonsense knowledge in the form of large amorphous semantic networks) from the work of John McCarthy, Allen Newell, Herbert A. Simon, Robert Kowalski and others whose work was based on logic and formal extensions of logic.[1] Schank described himself as an AI scruffy.[2] He made this distinction in linguistics, arguing strongly against Chomsky's view of language.[2][3]
The distinction was also partly geographical and cultural: "scruffy" attributes were exemplified by AI research at MIT under Marvin Minsky in the 1970s. The laboratory was famously "freewheeling" and researchers often developed AI programs by spending long hours fine-tuning programs until they showed the required behavior. Important and influential "scruffy" programs developed at MIT included Joseph Weizenbaum's ELIZA, which behaved as if it spoke English, without any formal knowledge at all, and Terry Winograd's[4] SHRDLU, which could successfully answer queries and carry out actions in a simplified world consisting of blocks and a robot arm.[5] SHRDLU, while successful, could not be scaled up into a useful natural language processing system, because it lacked a structured design. Maintaining a larger version of the program proved to be impossible, i.e. it was too scruffy to be extended.
Other AI laboratories (of which the largest were Stanford, Carnegie Mellon University and the University of Edinburgh) focused on logic and formal problem solving as a basis for AI. These institutions supported the work of John McCarthy, Herbert Simon, Allen Newell, Donald Michie, Robert Kowalski, and other "neats".
The contrast between MIT's approach and other laboratories was also described as a "procedural/declarative distinction". Programs like SHRDLU were designed as agents that carried out actions. They executed "procedures". Other programs were designed as inference engines that manipulated formal statements (or "declarations") about the world and translated these manipulations into actions. In his 1983 presidential address to Association for the Advancement of Artificial Intelligence, Nils Nilsson discussed the issue, arguing that "the field needed both". He wrote "much of the knowledge we want our programs to have can and should be represented declaratively in some kind of declarative, logic-like formalism. Ad hoc structures have their place, but most of these come from the domain itself." Alex P. Pentland and Martin Fischler of MIT concurred about the anticipated role of deduction and logic-like formalisms in future AI research, but not to the extent that Nilsson described.[6]
The scruffy approach was applied to robotics by Rodney Brooks in the mid-1980s. He advocated building robots that were, as he put it, Fast, Cheap and Out of Control, the title of a 1989 paper co-authored with Anita Flynn. Unlike earlier robots such as Shakey or the Stanford cart, they did not build up representations of the world by analyzing visual information with algorithms drawn from mathematical machine learning techniques, and they did not plan their actions using formalizations based on logic, such as the 'Planner' language. They simply reacted to their sensors in a way that tended to help them survive and move.[7]
Douglas Lenat's Cyc project was initiated in 1984 one of earliest and most ambitious projects to capture all of human knowledge in machine readable form, is "a determinedly scruffy enterprise".[8] The Cyc database contains millions of facts about all the complexities of the world, each of which must be entered one at a time, by knowledge engineers. Each of these entries is an ad hoc addition to the intelligence of the system. While there may be a "neat" solution to the problem of commonsense knowledge (such as machine learning algorithms with natural language processing that could study the text available over the internet), no such project has yet been successful.
As of 1991, Cyc contributor Marvin Minsky was still publishing papers evaluating the relative advantages of the neat versus scruffy approaches, e.g. “Logical Versus Analogical or Symbolic Versus Connectionist or Neat Versus Scruffy”.[9]
New statistical and mathematical approaches to AI were developed in the 1990s, using highly developed formalisms such as Bayesian nets and mathematical optimization. This general trend towards more formal methods in AI is described as "the victory of the neats" by Peter Norvig and Stuart Russell.[10] Neat solutions have been highly successful in the 21st century and are now used throughout the technology industry. These solutions, however, have mostly been applied to specific problems with specific solutions, and the problem of Artificial general intelligence (AGI) remains unsolved.
The terms "neat" and "scruffy" are not often used by AI researchers in the 21st century, with some exceptions such as talks by Karl Friston on the Free energy principle, where he refers to physicists as "Neats" and AI researchers as "Scruffies" (and philosophers as "Mystics"). "Neat" solutions to problems such as machine learning and computer vision are in widespread use,[10] but ad-hoc and detailed solutions still dominate research into robotics and commonsense knowledge.
Well-known examples
Neats
Scruffies
See also
Notes
- ^ Crevier 1993, p. 168
- ^ a b Brockman, John (7 May 1996). "9: Information is Surprises". Third Culture: Beyond the Scientific Revolution. Simon and Schuster. Retrieved 2 August 2021.
Chomsky has always adopted the physicist's philosophy of science, which is that you have hypotheses you check out, and that you could be wrong. This is absolutely antithetical to the AI philosophy of science, which is much more like the way a biologist looks at the world. The biologist's philosophy of science says that human beings are what they are, you find what you find, you try to understand it, categorize it, name it, and organize it. If you build a model and it doesn't work quite right, you have to fix it. It's much more of a "discovery" view of the world.
- ^ "Chapter 9 "INFORMATION IS SURPRISES"". www.edge.org. Retrieved 2 August 2021.
- ^ Winograd also became a critic of early approaches to AI as well, arguing that intelligent machines could not be built using formal symbols exclusively, but required embodied cognition. (Winograd 1986)
- ^ Crevier 1993, pp. 84−102, Russell & Norvig 2003, p. 19
- ^ Pentland and Fischler 1983, quoted in McCorduck 2004, pp. 421–424
- ^ McCorduck 2004, pp. 454–459
- ^ McCorduck 2004, p. 489
- ^ Wendy C . Lehnert (1 May 1994). "5: Cognition, Computers, and Car Bombs: How Yale Prepared Me for the 90's". In Schank, Robert; Langer, Ellen (eds.). Beliefs, Reasoning, and Decision Making: Psycho-Logic in Honor of Bob Abelson (First ed.). New York, NY: Taylor & Francis Group. p. 150. Retrieved 2 August 2021.
- ^ a b Russell & Norvig 2003, p. 25−26
References
- Crevier, Daniel (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.
- McCorduck, Pamela (2004), Machines Who Think (2nd ed.), Natick, Massachusetts: A. K. Peters, ISBN 1-5688-1205-1.
- Russell, Stuart J.; Norvig, Peter (2003), Artificial Intelligence: A Modern Approach (2nd ed.), Upper Saddle River, New Jersey: Prentice Hall, ISBN 0-13-790395-2.
- Winston, Patrick (1992). Artificial Intelligence. Addison Wesley. ISBN 978-0-201-53377-4.
- Gigerenzer, Gerd; Todd, Peter M.; et al. (ABC Research Group) (1999). Simple Heuristics That Make Us Smart. Oxford University Press. ISBN 9780199729241.
Further reading
- Anderson, John R. (2005). "Human symbol manipulation within an integrated cognitive architecture". Cognitive Science. 29 (3): 313–341. doi:10.1207/s15516709cog0000_22. PMID 21702777.
- Brooks, Rodney A. (2001-01-18). "The Relationship Between Matter and Life". Nature. 409 (6818): 409–411. doi:10.1038/35053196. PMID 11201756. S2CID 4430614.
- This article is based on material taken from the Free On-line Dictionary of Computing prior to 1 November 2008 and incorporated under the "relicensing" terms of the GFDL, version 1.3 or later.