Talk:History of artificial intelligence: Difference between revisions

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
Content deleted Content added
GreenC bot (talk | contribs)
Add 1 {{reflist-talk}} (via reftalk bot)
→‎Google, 2022: put in blurb from KBE which represents value incrementally accruing over time
Line 209: Line 209:
==Google, 2022==
==Google, 2022==
In 2022, Google is told to Havel developed an AI that is 158 million times faster than the world's fasterst supercomputer (source: [https://medium.com/predict/googles-quantum-computer-is-about-158-million-times-faster-than-the-world-s-fastest-supercomputer-36df56747f7f#:~:text=This%20makes%20Google's%20quantum%20computer,photons%2C%20electrons%20and%20atomic%20nuclei. Medium.com]) <!-- Template:Unsigned IP --><small class="autosigned">—&nbsp;Preceding [[Wikipedia:Signatures|unsigned]] comment added by [[Special:Contributions/151.38.135.105|151.38.135.105]] ([[User talk:151.38.135.105#top|talk]]) 20:39, 22 August 2022 (UTC)</small> <!--Autosigned by SineBot-->
In 2022, Google is told to Havel developed an AI that is 158 million times faster than the world's fasterst supercomputer (source: [https://medium.com/predict/googles-quantum-computer-is-about-158-million-times-faster-than-the-world-s-fastest-supercomputer-36df56747f7f#:~:text=This%20makes%20Google's%20quantum%20computer,photons%2C%20electrons%20and%20atomic%20nuclei. Medium.com]) <!-- Template:Unsigned IP --><small class="autosigned">—&nbsp;Preceding [[Wikipedia:Signatures|unsigned]] comment added by [[Special:Contributions/151.38.135.105|151.38.135.105]] ([[User talk:151.38.135.105#top|talk]]) 20:39, 22 August 2022 (UTC)</small> <!--Autosigned by SineBot-->

== Knowledge Base Engineering ==
Which is what we called it. Now, they're using Knowledge-based Engineering. But, it represents that each boom/(supposed) bust cycle left something of value. [http://%5B%5BKnowledge-based_engineering%5D%5D KBE] supported one large program meet demands of a new aircraft through all of its phases. The results were so impressive that subsequent programs adapted to the process as it morphed which is to be expected with computational systems. From a Lisp machine to Unix and then the PC, we can trace the evolution to a domain which still exists. We need to pull together this phenomenal reality. ... This motivated by looking at papers from a KSU Conference in 1986 (30 years after Darmouth) that had representatives from every effort extant at the time including those who brought about KBE. [[User:JMSwtlk|jmswtlk]] ([[User talk:JMSwtlk|talk]]) 14:14, 27 January 2023 (UTC)

Revision as of 14:14, 27 January 2023

Good articleHistory of artificial intelligence has been listed as one of the Engineering and technology good articles under the good article criteria. If you can improve it further, please do so. If it no longer meets these criteria, you can reassess it.
Article milestones
DateProcessResult
September 28, 2007Peer reviewReviewed
October 18, 2008Good article nomineeListed
Current status: Good article

Template:Vital article

WikiProject iconHistory of Science GA‑class Mid‑importance
WikiProject iconThis article is part of the History of Science WikiProject, an attempt to improve and organize the history of science content on Wikipedia. If you would like to participate, you can edit the article attached to this page, or visit the project page, where you can join the project and/or contribute to the discussion. You can also help with the History of Science Collaboration of the Month.
GAThis article has been rated as GA-class on Wikipedia's content assessment scale.
MidThis article has been rated as Mid-importance on the project's importance scale.
WikiProject iconComputing GA‑class High‑importance
WikiProject iconThis article is within the scope of WikiProject Computing, a collaborative effort to improve the coverage of computers, computing, and information technology on Wikipedia. If you would like to participate, please visit the project page, where you can join the discussion and see a list of open tasks.
GAThis article has been rated as GA-class on Wikipedia's content assessment scale.
HighThis article has been rated as High-importance on the project's importance scale.

Wiki Education Foundation-supported course assignment

This article is or was the subject of a Wiki Education Foundation-supported course assignment. Further details are available on the course page. Student editor(s): VjiaoBlack.

Above undated message substituted from Template:Dashboard.wikiedu.org assignment by PrimeBOT (talk) 23:35, 16 January 2022 (UTC)[reply]

Turing tradition

The article should perhaps mention the Turing tradition which is perhaps less well known on the other side of the pond. The Turing tradition is an approach to machine learning based on "(i) the use of logic and (ii) close attention to practical problems". [1] It is a common theme in the work of Alan Turing, Donald Michie, Ehud Shapiro, Ross Quinlan, Stephen Muggleton. Pgr94 (talk) 08:43, 16 August 2008 (UTC)[reply]

The school of thought is certainly influential in the development of Inductive Logic Programming and probably also Abductive Logic Programming. Pgr94 (talk) 09:01, 16 August 2008 (UTC)[reply]
If I'm reading this right, this is a tradition (1) popular in England (2) based on logic (3) focussed on machine learning, right? Do you think it's fair to throw this in with other logical approaches to AI in England, i.e. Robert Kowalski at Edinburgh University, etc.? ---- CharlesGillingham (talk) 00:47, 17 August 2008 (UTC)[reply]
I agree this should be covered. I've had to dig for more material about England, Europe and Japan. Just one sentence under "Logic" should do it. This may also belong in AI#Traditional symbolic AI#Logical AI and probably in machine learning#History (if someone gets around to writing a History section for that article). The key question is, how influential was it? I'd like to find a second source. ---- CharlesGillingham (talk) 01:01, 17 August 2008 (UTC)[reply]
Schools of thought don't respect national boundaries, so "popular in England" is not how I'd describe it. ILP is significant on an international level (although hasn't received much attention in the US for some reason). Michie was a colleague of Turing; Muggleton (principle player in ILP) was his student. The influence of Turing is clearly there. Gillies named it, but perhaps the few refs suggest that the name is not that notable[2] [3] Pgr94 (talk) 11:08, 18 August 2008 (UTC)[reply]
A small point perhaps but the UK is not England and Edinburgh is in Scotland, also part of the UK. 17:06, 7 March 2015 (UTC) 86.9.223.140 (talk)

A few gaps

I was surprised to see no mention of:

  • Chess. Claude Shannon wrote about it, Mikhail Botvinnik made a little progress in develpoing an algorithm that "thought" like a human player (these articles contain refs). See also Computers and Chess, which may provide WP:RS refs.
  • Computer game AIs.
  • Incorporation of AI-like features into e.g. spam-filters in email programs, many of which respond to "training".

I also remember someone writing (? in the 1990s) that, every time an AI research field looked like it was getting somewhere, the goalposts were moved - e.g. when chess programs became competitive, AI was redefined to exclude them and focus on e.g. visual recognition. -- Philcha (talk) 17:10, 5 October 2008 (UTC)[reply]

Keeping in mind that this article is already a bit long, we could cover these topics:
  • Computer chess and computer checkers. A paragraph describing the experiments of Claude Shannon, Christopher Strachey and Arthur Samuel could be added between "Turing's test" and "Logic theorist" sections. This was an early test bed for AI research, especially in the 40s and 50s. McCorduck devotes a chapter to this topic. Crevier does as well. Russell and Norvig mention it. (There was such a paragraph in early drafts of the article, but I cut it for length. I felt that it was less important than logic, natural language, microworlds, connectionism, etc.. I'm open to arguments to the contrary.)
  • Game AI I think this could be mentioned as one of the list of applications in AI behind the scenes. I'm not sure if there is more to say about it than that. Is it more historically influential than that? (I don't have a good reference for this).
  • Spam filtering could be added to list the successful applications mentioned in AI behind the scenes as well. Perhaps like this: "spam filtering (which uses sophisticated machine learning algorithms)"
  • AI effect. This is described in the second paragraph of AI behind the scenes. (The article on this topic was deleted for non-notability. A new draft is at User:CharlesGillingham/AI effect.)
---- CharlesGillingham (talk) 23:53, 5 October 2008 (UTC)[reply]
Congratulations on the GA rating!
CharlesGillingham, I've looked at your User:CharlesGillingham/AI effect and all you need to do to prove notability is put a few of the refs inline.
A bit more about Computer game AIs - this is where non-specialist readers may be most familiar with the phrase "AI" (the other is in quasi-science fiction movies). Game AIs illustrate a lot of the difficulties, e.g. they are poor at path-finding (getting a group of units from A to B without traffic jams and without some units wandering off and getting isolated) and at planning (they rely on pre-scripted economic development and attack sequences, and on various "unfair" advantages). WP:RS for this would mostly be at developer mags, e.g. Gamasutra. The best game AIs I'm aware of are Total Annihilation (from personal observation, units act as teams - if they have more firepower than is needed to destroy the designated target, they do a good job of picking a secondary target and deciding who should shoot what) and Galactic Civilizations (praised for its good planning and clever strategies, apparently without cheating).
You might need just one more sentence about the impact of Moore's law for the benefit of non-specialists - e.g. the mid-range desktop on which I'm writing this is over 200 times faster than IBM's top mainframes were in the mid-1970s and has 1000x more RAM, and the first researchers where using 1950s machines with less processing power than a modern digital watch.
The Tesler quote (1970) about "AI" being whatever computers can't do yet would fit well into "AI behind the scenes".
Section "Nouvelle AI" could do with a sentence summmarising 1 or 2 projects based on synthetic animals. -- Philcha (talk) 13:38, 20 October 2008 (UTC)[reply]

All this Kurzweil Self-Promotion

I found five references to Ray Kurzweil, who did not contribute anything to Artificial Intelligence! Apparently inserted by Kurzweil himself (or perhaps by somebody in his company's PR department). In the field of AI he certainly is not regarded as an influential researcher, although he writes a lot about the future of AI, elaborating on big ideas introduced by others, such as the technological singularity popularized by Vernor Vinge 20 years ago. I suggest to remove those Kurzweil references, and focus on people who really had an impact on AI history. Quiname (talk) 20:42, 14 April 2011 (UTC)[reply]

Kurzweil is used in the text only as an example of current optimism. I think this is appropriate and I can't think of anyone more optimistic or more popular. Kurzweil's Singularity is also used to cite a few things (such as the "AI effect"). This is appropriate, I think, because the book is a popular introduction to AI, and popular introductions (i.e. WP:SECONDARY sources) are good sources for an article such as this, which is partly about AI, but also partly about what people say about AI. ---- CharlesGillingham (talk) 22:56, 14 April 2011 (UTC)[reply]

Tone of the Introduction

The tone (WP:TONE, WP:PEACOCK) of this article's introduction is too triumphant and embellished. I don't know enough about the topic to improve it but it could undoubtedly use improving. Exercisephys (talk) 16:45, 25 May 2013 (UTC)[reply]

AI used for long duration Space Exploration

Consider developing AI for space exploration. AI would have the ability to make decisions on flight path, speed and modifications of vehicle to enhance speed, data transmission to point of origin, data processing and storage among other goals. It could have constant communication with a terra-based AI system for transmission of knowledge and reporting.

Would be best possible method for humans to explore at least the local cluster of stars at 0.5 to 0.9 light speed travel.

unsigned comment added by 70.43.17.173 (talk) 20:47, 29 October 2015 (UTC) TonSerra (talk) 15:59, 30 October 2015 (UTC)[reply]

No longer up-to-date

This article doesn't cover the last 15 years, when AI (and especially deep learning) have begun to dominate finance, information technology, science, and industry. I don't plan on writing this, but if I did, the topics would be (1) deep learning: big data + fast machines + statistical AI == ginormous success. (2) artificial general intelligence, which has coalesced into a new and interesting subfield, with a very high profile, very high hopes and no real successes. Brain simulation (Numenta, Blue Brain) deserves a mention here. (3) Jaron Lanier and Noam Chomsky's criticisms of deep learning and statistical AI in general: they argue it's not AGI, it's really just statistics. ---- CharlesGillingham (talk) 06:28, 10 September 2016 (UTC)[reply]

I realize you wrote this 5 years ago, although it remains mostly reasonable. I'm not sure if the article has been updated in the interim or not. Also, I'm not sure at what point in time we should consider the line of demarcation between historical artificial intelligence and the current full-blooming artificial intelligence of the present in 2021... such as it is. I'm not replying in order to dump on AI despite my last sentence. I think, with the benefit of 5 year's hindsight, I would break out an update of the article to include:
1 machine learning which is a term used synonymously with AI sometimes. For anti-money laundering, financial risk management, fraud/abuse detection in general (e.g. analyzing computer and network log files to surface anomalies and pattern matches), e-commerce, and law enforcement-related surveillance applications, yes, machine learning and the data/database architectures and appliances that are often used to implement them at scale (e.g. Hadoop, Hive, Pig, MongoDB, Kubernetes, Netezza), are successes. As for deep learning (versus ML), maybe, but successes haven't been nearly as well documented.
2 artificial general intelligence ditto, with the addition of Neuralink and maybe others
3 Criticisms of machine learning AND deep learning for not being AGI: yes, but there are subject matter experts that are more credible than Jaron Lanier for that purpose, I believe. (I have no problem with Jaron Lanier as you can read here if you wish; jeez louise, I wrote that back in 2012). Seems better to leave Noam Chomsky out of criticizing anything on the basis of it being merely statistics too.
I *am* curious (as a sometimes statistician, often applier of probability theory IRL) about the specifics of Noam Chomsky's critique of deep learning as just statistics, CharlesGillingham if you wouldn't mind elaborating a bit.--FeralOink (talk) 08:22, 12 July 2021 (UTC)[reply]
These points have been added to the article (See the final section of the article). Critiques of AI's current approach are not mentioned, and I think that makes sense. We'll add a paragraph about that after Statistical AI fails, during the next AI winter ;).
I don't think we need to branch the topic of "AI" because, well, we still call it by the same name, and I think the whole point here is to see all the ways AI has tried and failed in the past. The issues that came up years ago are issues AI will need to deal with in the future. ---- CharlesGillingham (talk) 18:51, 12 July 2021 (UTC)[reply]

External links modified

Hello fellow Wikipedians,

I have just modified 3 external links on History of artificial intelligence. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:

When you have finished reviewing my changes, you may follow the instructions on the template below to fix any issues with the URLs.

This message was posted before February 2018. After February 2018, "External links modified" talk page sections are no longer generated or monitored by InternetArchiveBot. No special action is required regarding these talk page notices, other than regular verification using the archive tool instructions below. Editors have permission to delete these "External links modified" talk page sections if they want to de-clutter talk pages, but see the RfC before doing mass systematic removals. This message is updated dynamically through the template {{source check}} (last update: 18 January 2022).

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—InternetArchiveBot (Report bug) 09:49, 3 April 2017 (UTC)[reply]

External links modified

Hello fellow Wikipedians,

I have just modified one external link on History of artificial intelligence. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:

When you have finished reviewing my changes, you may follow the instructions on the template below to fix any issues with the URLs.

This message was posted before February 2018. After February 2018, "External links modified" talk page sections are no longer generated or monitored by InternetArchiveBot. No special action is required regarding these talk page notices, other than regular verification using the archive tool instructions below. Editors have permission to delete these "External links modified" talk page sections if they want to de-clutter talk pages, but see the RfC before doing mass systematic removals. This message is updated dynamically through the template {{source check}} (last update: 18 January 2022).

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—InternetArchiveBot (Report bug) 01:16, 5 November 2017 (UTC)[reply]

Do we really need Pamela McCorduck in the 1st paragraph? Or at all?

Looks like promotion to me. — Preceding unsigned comment added by Callmesolis (talkcontribs) 22:12, 9 February 2019 (UTC)[reply]

Vague "weasel" language?

One sentence struck me in this article, regarding the vanishing gradient problem. Overcoming this problem was central to the advance of Deep Learning, and Long short-term memory models were the first and still are one of the most widely used models that resolved this critical issue. So the current text "There have been many methods developed to approach this problem, such as Long short-term memory units." seems to play down the important role of these types of networks, while being vague in not listing any other of the "many methods".

Unless someone wants to suggest some other prominent solutions to the vanishing gradient problem that I am not aware of (more modern ones, perhaps?) I would propose revising this section to clarify that LSTMs were a key breakthrough in the field, precisely because they were the first solution to the vanishing gradient problem, and still to date constitute a prevalent network type in modern deep learning research and applications. — Preceding unsigned comment added by 62.178.202.229 (talk) 14:19, 13 July 2019 (UTC)[reply]

Expansion / Reorganization of "Precursors"

I plan on restructuring the "precursors" section by subdividing it into two sub-sections: "Mythical, Fictional, and Speculative" and "Theoretical and Technological." The bulk of what is currently in the precursors section will go under the second heading. I plan on expanding upon the first heading significantly. (Not more than ~500 words.) HieronymusBot (talk) 18:17, 15 March 2020 (UTC)[reply]

Neats vs scruffies

Every once in a while, during the past 15 years, someone comes through a deletes all references the terms "neat" and "scruffy". So let's talk about it.

I've put the terms back into the article because: (a) "neats vs. scruffies" is genuine, real history: people talked about this (a lot) in the late 70s and 80s. See AAAI conferences and talks and so on. Part of writing history is describing the world as it was, as it described itself. They described themselves this way, so we need to give the reader an insight into how they thought at the time. (b) Russell and Norvig's "victory of the neats" quip provides a nice through-line for the article, and emphasizes the fact that each generation has viewed the field differently as it has evolved. This is also good history writing, and there is simply is no more reliable source on AI than Russell and Norvig.

On the other side, by the late eighties, most people were sick of hearing about it, and sick of the ridiculous ways that people would try to weave the distinction into badly thought out "general theories" of AI or cognition. I remember Robert Wilensky telling me in 1987: "Never read anything that mentions neats vs scruffies or procedural vs. declarative". In short: we realized it was a stupid distinction years later, but that doesn't mean it isn't historically relevant.

If your disagree, please, let me know what the actual problem is. ----CharlesGillingham (talk) 16:44, 5 July 2020 (UTC)[reply]

"Neats and scruffies" was not as prevalent in the community as this WIKI entry makes it out to be. It was limited to a small group of admittedly combative researchers, who did not typically describe themselves that way--but did describe their processes in that manner. Utilizing a process does not make one a particular type of researcher or developer, and thus limits the application of the term. The term doesn't show up in any relevant reporting of the day (aka, Freedman, Newquist, Levy, any computer journals). It can be referenced, perhaps to another entry, but to use that term as a linchpin for this entry is completely disproportionate to its use then and now. I would like to see an actual set of citations from the era describing the developers in actual terms such as "XXX is a scruffy" or "XXX and the team of neats." Thanks. TrainTracking1 (talk) 20:31, 2 April 2021 (UTC).[reply]
TrainTracking1, here is a reference: Minsky, Marvin L. "Logical versus analogical or symbolic versus connectionist or neat versus scruffy." AI magazine 12.2 (1991): 34-34. [4]. --Hectorpal (talk) 02:13, 1 August 2021 (UTC)[reply]
Weighing in. Completely against the isolated use of "neats vs scruffies." It's what the real world calls "inside baseball"--as far as that goes--in that it has no relevance to the greater discussion, and is not at all relative to the understanding of the history of AI. You might has well start talking about McCarthyists vs Minskyites if you want to get into conflict. And by 1991, the terms were already out of whatever fashion they were in. Perhaps it's time for someone to start a "neats vs scruffies" page, which should include the use of the term in other disciplines, as it apparently is not limited to AI.Andreldritch (talk) 22:37, 1 August 2021 (UTC)[reply]

Wrong attribution of ada lovelace quotation

Ada Lovelace is quoted: "might compose elaborate and scientific pieces of music of any degree of complexity or extent", however in the linked source (https://johnrhudson.me.uk/computing/Menabrea_Sketch.pdf) this is only found in the section "notes by the translator". — Preceding unsigned comment added by Cashney (talkcontribs) 11:48, 9 August 2020 (UTC)[reply]

Please protect the "neats" and "scruffies"

I just read those terms are being deleted. Please protect them. It was very important. To be honest, this is going to come back. "neats" are challenging Deep Learning who are for them the new "scruffies". Just see Judea Pearl:neat to see this in progress. Hectorpal (talk) 16:36, 1 April 2021 (UTC)[reply]

Neats and scruffies is only a term used by SOME in the AI community (historically, Schank is the only person to be associated with it). It is also a popular characterisation in many areas of academia, and is rarely used in the actual history of AI--except by those who consider themselves on one side or the other. Labeling diminishes the efforts of various schools and the wide variety of crossover. Suggest leaving the battle to the actual WIKI entry on Neats and Scruffies. Keep history entry about the actual events and not the insider squabbling. TrainTracking1 (talk) 20:22, 2 April 2021 (UTC)[reply]
Hectorpal, your link to archetypal neat URL is a 404. Did you mean this neat? --FeralOink (talk) 14:51, 8 July 2021 (UTC)BTW you have a great LinkedIn pic! TrainTracking1, as an uninvolved editor, I would keep the neats and scruffies and improve the sourcing to McDuck please see my comment below. I am tempted to PROD her BLP. This is what it consists of at the moment:[reply]

McCorduck grew up in California and attended the University of California, Berkeley, from which she graduated in 1960. McCorduck was invited to contribute to a book of readings on artificial intelligence while a senior at the University of California, Berkeley, in 1960. At the time she did not know what artificial intelligence was. McCorduck lived for more than forty years in New York City with her husband Joseph F. Traub. After her husband's death she moved back to California, where she had grown up. She now lies [sic] in San Francisco.

Word count for "McCorduck" is 74 on this article.--FeralOink (talk) 10:50, 8 July 2021 (UTC)[reply]
FeralOink, yes, thanks. Fixed. --Hectorpal (talk) 02:09, 1 August 2021 (UTC)[reply]

@TrainTracking1: Absolute incorrect on all counts. It was only used 1975-1985. It is mentioned in the history section leading AI textbook, Russell & Norvig. It is mentioned in both of the most popular and respected histories of AI, Crevier and McCorduck. It wasn't not unique to Schank (at the time, I didn't even know it was Schank who came up with it). It was the topic of talks and symposiums at AAAI. It was addressed in the presidential address of AAAI several times. Papers were written about it (usually including the "procedural/declarative distinction"). Also the rivalry between MIT and Stanford was very real -- each side thought the other was dead wrong.

I agree that, by 1985, everybody was sick of hearing about it. A lot of worthless papers were written about it, papers that were trying to establish some kind of new paradigm for the field or their own "better" definition of AI and so on. These kind of papers are useless, boring and a dime a dozen. People still write these kinds of papers today, e.g. "Defining 'Synthetic Consciousness'". They are just as useless now.

But that doesn't mean that scruffy/neat doesn't raise an interesting question about AI. Is there a simple and elegant "master algorithm" for AGI? Or do we necessarily have to solve a lot of messy unrelated problems? ---- CharlesGillingham (talk) 19:48, 8 July 2021 (UTC)[reply]

Forgive, I didn't realize we talked about this a few months ago. Forgive me for restating my position. ---- CharlesGillingham (talk) 19:50, 8 July 2021 (UTC)[reply]

Here is a reference on neat vs scruffy. Minsky, Marvin L. "Logical versus analogical or symbolic versus connectionist or neat versus scruffy." AI magazine 12.2 (1991): 34-34. [5]. 473 citations in Google Scholar. The expressions belong in the history of AI.<unsigned comment>

At least one mention of neats and scruffies seems justified, especially since there is an already extant Wikipedia article on the subject in the context of AI. I'll see if or where it might fit in this article. Maybe the above source will be useful, whomever deposited it here.--FeralOink (talk) 10:33, 2 August 2021 (UTC)[reply]
I am doing a major re-write and update of neats and scruffies as the sources aren't great, and as such, dates and context are missing in a lot of places. See [my revisions here to date], if anyone is curious.--FeralOink (talk) 14:06, 2 August 2021 (UTC)[reply]

There are 74 hits on McCorduck in this article

They are all in the sources, but that still seems like a lot. Better sources added incrementally over time might be good. I am only suggesting, and realize I can do that too. Help out, that is.--FeralOink (talk) 14:55, 8 July 2021 (UTC)[reply]

McCorduck wrote the definitive history of AI. There is no better source. The fact that there are so many references to it is an indication that this article uses only the most reliable sources -- no fringe points of view, no random semi-relevant contributions. Just the mainstream, consensus understanding of the history of AI.
I would argue that articles about established academic topics (such as history) are more likely to reliable if they depend on less sources, not more -- or even just one: the most respected mainstream source. The other random sources tend to be about topics that are either unimportant, fringe, or (at worst) self-serving. ---- CharlesGillingham (talk) 19:26, 8 July 2021 (UTC)[reply]
Completely disagree about McCorduck and "the definitive history." Her book was published in 1979, well before the commercial development of AI--and its attendant developers and corporate purveyors--was even established (ranging from Symbolics and LISP Machine to Intellicorp and Inference). In fact, her only association with the commercial rise of AI was her work with Feigenbaum, which left them both on the wrong side of the "Fifth Generation" call to arms. To state that the use of her as the most oft-cited source is indicative of nothing . . . other than familiarity with her work by an early Wiki editor on the topic of AI. (That's a fallacy akin to saying the existence of so many yellow cars in NYC is indicative of that being the best choice of colors for cars.) Other writers, like Norvig, Freedman, Newquist, and Crevier wrote about the rise of commercial AI in much more detail than McCorduck--essentially because she barely touched on it at all. To cite her, or any single author, as the creator of the definitive history is simply misguided and biased. And, of course, leans into the bias of using only one predominant source. McCorduck is not the only highly-regarded chronicler of AI, nor should she (or anyone else) be given that title. Let's get some more sources in here. TrainTracking1 (talk) 06:17, 11 July 2021 (UTC)[reply]
Arguing in favor of historical articles being best if dependent on a single historical source is antithetical to a balanced point of view (unless for an historical article which occurred, say, prior to 400 B.C. and sources are scarce.) For History of AI, single-sourcing will almost certainly result in WP:UNDUE and WP:NPOV for the article overall. Thank you for the suggestions, TrainTracking1. I will try to find more written by Norvig, Freedman et al. I encourage other editors to do similarly. A great deal has happened in the history of AI that post-dates 1979 (publication of McCorduck's book), but is still part of the history—not the present—of artificial intelligence.--FeralOink (talk) 10:25, 2 August 2021 (UTC)[reply]

Qualifying the Victory of the Neats: Updating a Section Name from the "Victory of the Neats" to "Probabilistic Reasoning and Greater Rigor" and Updating the Text

I see there was a lot of discussion on this talk page about whether to preserve 'Neats vs Scruffies' or remove it.

The latest version Russell & Norvig's, AI: A Modern Approach differs from the second edition cited earlier by changing the last sentence of the footnote on P.25 of the 2nd edition and P. 24 of the 4th edition from "Whether that stability will be disrupted by a new scruffy idea is another question" to --now-- "The present emphasis on deep learning may represent a resurgence of the scruffies."

I think the new Russell & Norvig characterization there as and historical breakdown better describes that section, so I am changing the name to more closely match what they have. I'm also trying minimize disruption and flow of the article. I had planned on just dropping the part of the sentence "...and the victory of the neats" in Russell & Norvig (2003) describe this as nothing less than a "revolution" and "the victory of the neats".

But since I can see why others care about that the 'neats vs scruffies' view and possible future application, I am adding:

They had argued in their 2002 textbook that this increased rigor could be viewed plausibly as a "victory of the neats,"[1] but subsequently qualified that by saying, in their 2020 AI textbook, that "The present emphasis on deep learning may represent a resurgence of the scruffies."[2] Veritas Aeterna (talk) 23:30, 7 July 2022 (UTC)[reply]

References

  1. ^ Russell, Stuart J.; Norvig, Peter (2002-12-01). Artificial Intelligence: A Modern Approach (2nd edition ed.). Upper Saddle River, N.J: Prentice Hall. ISBN 978-0-13-790395-5. {{cite book}}: |edition= has extra text (help)
  2. ^ Russell, Stuart; Norvig, Peter (2020-04-28). Artificial Intelligence: A Modern Approach (4th edition ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. {{cite book}}: |edition= has extra text (help)

Google, 2022

In 2022, Google is told to Havel developed an AI that is 158 million times faster than the world's fasterst supercomputer (source: Medium.com) — Preceding unsigned comment added by 151.38.135.105 (talk) 20:39, 22 August 2022 (UTC)[reply]

Knowledge Base Engineering

Which is what we called it. Now, they're using Knowledge-based Engineering. But, it represents that each boom/(supposed) bust cycle left something of value. KBE supported one large program meet demands of a new aircraft through all of its phases. The results were so impressive that subsequent programs adapted to the process as it morphed which is to be expected with computational systems. From a Lisp machine to Unix and then the PC, we can trace the evolution to a domain which still exists. We need to pull together this phenomenal reality. ... This motivated by looking at papers from a KSU Conference in 1986 (30 years after Darmouth) that had representatives from every effort extant at the time including those who brought about KBE. jmswtlk (talk) 14:14, 27 January 2023 (UTC)[reply]