Wikipedia talk:Bots/Requests for approval: Difference between revisions

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
Content deleted Content added
→‎Arbitrary break: of course BAG is against unapproved bots
Line 220: Line 220:
::::::::The BAG doesn't have a magical ability to prevent someone form running an unapproved bot. There is no technical limitation that the BAG enables to allow a bot to run. --[[User:ThaddeusB|ThaddeusB]] ([[User talk:ThaddeusB|talk]]) 19:28, 27 June 2009 (UTC)
::::::::The BAG doesn't have a magical ability to prevent someone form running an unapproved bot. There is no technical limitation that the BAG enables to allow a bot to run. --[[User:ThaddeusB|ThaddeusB]] ([[User talk:ThaddeusB|talk]]) 19:28, 27 June 2009 (UTC)
:::::::::Yet there are no consequences to running an unapproved bot. A post about an unapproved bot was ignored. It's not even interesting that someone runs an unapproved bot. --[[Special:Contributions/69.226.103.13|69.226.103.13]] ([[User talk:69.226.103.13|talk]]) 20:09, 27 June 2009 (UTC)
:::::::::Yet there are no consequences to running an unapproved bot. A post about an unapproved bot was ignored. It's not even interesting that someone runs an unapproved bot. --[[Special:Contributions/69.226.103.13|69.226.103.13]] ([[User talk:69.226.103.13|talk]]) 20:09, 27 June 2009 (UTC)
::::::::::Your assertion that BAG members would ignore an unapproved bot is absurd. The bot that caused these problems was an approved bot. You fundamentally misunderstand the difference between changing an aspect of a bot's code and a completely new task. Of course someone doesn't have to seek approval every time they modify their code. Bots are approved for a ''task'' and it is perfectly permissible to change one's code to better do the same task without seeking new approval.
::::::::::Also, many users have been banned from Wikipedia or otherwise disciplined for running unapproved bot, so there certainly is not "no consequence" as you assert. --[[User:ThaddeusB|ThaddeusB]] ([[User talk:ThaddeusB|talk]]) 21:00, 27 June 2009 (UTC)


:::::::It was me who suggested that. I actually thought that I'd proposed noindexing the articles earlier this week - but looking back at my contribs, this seems not to be the case (perhaps I got edit conflicted? dunno). Unless this is prohibited by policy somewhere, this would be a Very Good Idea, IMO. 69.226. is correct in saying that these broken articles don't exactly make WP look good... --[[User:Kurt Shaped Box|Kurt Shaped Box]] ([[User talk:Kurt Shaped Box|talk]]) 18:18, 27 June 2009 (UTC)
:::::::It was me who suggested that. I actually thought that I'd proposed noindexing the articles earlier this week - but looking back at my contribs, this seems not to be the case (perhaps I got edit conflicted? dunno). Unless this is prohibited by policy somewhere, this would be a Very Good Idea, IMO. 69.226. is correct in saying that these broken articles don't exactly make WP look good... --[[User:Kurt Shaped Box|Kurt Shaped Box]] ([[User talk:Kurt Shaped Box|talk]]) 18:18, 27 June 2009 (UTC)

Revision as of 21:00, 27 June 2009

"Deapproval"

Could someone please tell me what the process is (if indeed there is a process) for getting the approval of a bot task overturned? ListasBot's third task was approved after a discussion involving only two editors (the bot owner and one BAG member), and I feel the the decision may have to be reassessed. Thanks, — Martin (MSGJ · talk) 15:05, 9 May 2009 (UTC)[reply]

Block it, and then create discussion somewhere - an RFC, talk page, or VPM would be fine. - Jarry1250 (t, c) 16:08, 9 May 2009 (UTC)[reply]
Actually, the official "process" at this time is to begin a discussion on this talk page. Immediate blocking would IMO not be appropriate in many cases; that should only be done when the bot is actively causing harm or acting against consensus and not just because any random admin wants to start a discussion. Anomie 21:18, 9 May 2009 (UTC)[reply]
Sorry, I was assuming harm was being done. No, you're quite right, better to discuss first. - Jarry1250 (t, c) 21:27, 9 May 2009 (UTC)[reply]

FYI, I've moved this discussion here. Thanks, Matt (talk) 03:25, 12 May 2009 (UTC)[reply]

Reexamination of Tomerbot

Unfortunately it was decided not to approve my Bot to its main purpose which is making categories out of lists. I would still need a bot flag for the task of putting Hebrew Interwikis so the bot can operate correctly in the Hebrew Wikipedia. Tomer A. 13:58, 15 May 2009 (UTC)[reply]

You may file another BRFA as "Tomerbot 2" if you just want approval to do IW. –xeno talk 14:05, 15 May 2009 (UTC)[reply]
Yeah, that should be quick and uncontroversial. – Quadell (talk) 14:45, 15 May 2009 (UTC)[reply]
10x. Tomer A. 20:08, 17 May 2009 (UTC)[reply]

a serious problem

A number of bots rely on the {{DEFAULTSORT}} or {{Persondata}} parameters, or the listas parameter in biography templates. The data encoded in these templates may be reliable for individuals whose names are in the European naming style of inherited surnames.

Unfortunately, over the years, well-meaning, but ill-advised volunteers have mistakenly assumed Arabic and Chinese names should also be shoehorned into the European naming scheme. This is a problem. And bots that treat this data as reliable are compounding an already serious problem.

I suggest no bot that relies on this unreliable data should be approved. Geo Swan (talk) 08:17, 27 May 2009 (UTC)[reply]

People reading this should probably look at WP:Administrator's noticeboard/Incidents#A rogue bot (most recent diff) and User_talk:Mikaey#It is a big mistake to try to shoehorn traditional Arabic names into the European naming scheme of inherited surnames... to get an idea of what's going on. Matt (talk) 08:26, 27 May 2009 (UTC)[reply]
The Defaultsortbot is only one of the bots recently approved that compounds this mistake. The recently approved listasbot also relies on the unreliable data in these parameters. Various robot-assisted editing tools rely on the unreliable data in these parameters. Geo Swan (talk) 09:14, 27 May 2009 (UTC)[reply]
As you have been told several times already in the linked discussions, the bot is not adding DEFAULTSORT de novo; it is only copying what a human editor has already placed in a listas parameter, persondata template, or the like. There is not much that can be done about GIGO, so unless you can provide evidence that this problem is widespread enough that a significant fraction of the bot's edits will be erroneous I can't see that anything needs to be done here. One possibility, if you can supply a sufficiently accurate heuristic, would be for the bot to ignore or log pages with seemingly-Arabic names for human review.
BTW, Geo Swan, I notice in your contribs that you recently went through a number of articles on people with Arabic names and removed the listas parameter from {{WPBiography}}. It would have made much more sense to replace it with the "correct" value instead, possibly also with a comment pointing out why that is the correct value. Otherwise, well-meaning but misinformed editors are likely to repeat the same mistake. Anomie 12:26, 27 May 2009 (UTC)[reply]
Let be clear about impatience my raising of the corrupt information the bots are relying on has generated. Don't shoot the messenger. This situation is an enormous mess, and it is a mess I do not believe I played any role in generating. I regard this impatience as inappropriate. I would appreciate those expressing it to reserve any expressions of impatience until after we have discussed the various suggestions I have been getting about what I did wrong.
I know I have been told that the various bots and humans assisted by robot editing tools are not naively shoehorning bogus surnames into articles about non-Europeans "de novo". And I know this is incorrect. Here is an instance:
If you look at the edit summary that includes "AWB", it means that it was actually Wikipedia:AutoWikiBrowser that suggested that, in the attempt to add the "listas" section to the template. I didn't know it was incorrect, so I let the tool do its thing.
User:Raven1977 was responding to a question I asked him about his or her addition of a bogus listas to Talk:Abbas Habid Rumi Al Naely. A few days earlier another contributor had added a bogus listas to Talk:Abbas Habid Rumi Al Naely. I asked him or her why. And that first contributor told me their edit was based on the (bogus IMO) value in the article's {{DEFAULTSORT}}. I subsequently removed the article's {{DEFAULTSORT}}. So this second guy, with the assistance of AWB, was naively adding these bogus values "de novo". I corrected the articles this second guy was naively munged, that were on my watchlist. When I looked at his most recent contribution history I saw he or she had recently added listas fields to SEVERAL HUNDRED other articles on individuals with Arabic names.
Their comment, "...so I let the tool do its thing." As you stated above, a human should be in the loop. A human should be making an informed choice as to whether to follow, or over-ride, the bot's suggestion. Alarmingly Raven1977 was abdicating responsibility to provide the informed decision-making. I wrote the author of AWB. He told me that AWB only suggested inserting a default sort, or listas, when a human had already made an informed choice that one was appropriate. So, how come Raven1977's use AWB was suggesting bogus values "de novo"? I dunno.
I am afraid you are mistaken. Early in the history of placing DEFAULTSORT and listas parameters there were bots that went around placing wild guesses at what individuals surnames were, without consideration to the billions of individuals who had names that didn't fit in that scheme. User:Jim Cubb acknowledged this on my talk page, and I have come across records of these bot's operations myself. So, no offense, I do believe a "significant fraction" of the bot's operation will be erroneous.
As I tried to point out to Raven1977, when he or she was compounding the problem by relying on unreliable advice from AWB, they were affecting 3 articles per minute. In my limited attempts to clean up after Raven1977 I was averaging about one article every three minutes. See the problem? The clean-up is much more time-consuming than compounding the error in the first place.
Anomie, implicit in your comment is the assumption that treating individuals with European style inherited lastname-surnames should be considered the standard case. I think that is backwards. Billions of individuals have names best sorted starting with the first character, proceeding smoothly to the last character. Arabic influenced names aren't the only ones. 1.3 billion Chinese people have surnames -- but their surnames begin their name, not end it. It is European style names that need to be treated as the exception -- not the default. In my opinion these bots should be leaving alone all biographical articles unless there is some reason to believe the individual's name uses the European style.
Various other people have suggested that, instead of removing the bogus listas and DEFAULTSORT values, I should have replaced them with "correct" values. No one has asked if I think this is a good idea. Actually, I considered doing this. I considered it, and decided it was a really bad idea. I have about 700 1000 articles about individuals with Arabic names on my watchlist. The more references I come across to these individuals, the more choices I see to transliterations and renderings we would need to chose among to pick a base name the other names should be redirected to. Most of those 1000 names have been moved, at least once. When I choose, or participate in the choice of which alternative transliteration should be at the base name, the choice is largely arbitrary. And it is likely to be reversed, or superceded.
It is safe assumption that articles about individuals with Arabic names are likely to be renamed. So, adding in the current article name, in the listas field, and in the DEFAULTSORT parameter, is simply an invitation to an additional maintenance burden. In my opinion, it is a totally unnecessary one. In my opinion it should be the European-styled names that should be considered the exception. Non-European-styled names should be able exist without a mutable, troublesome defaultsort, when the sort key for them is identical to whatever the current name of the article is, not the name that was current when the bogus Europeanized surname was shoehorned in.
Let me be frank. Whoever first thought up the idea of automating sorting, made a mistake when they didn't foresee that the billions of individuals who don't have European style inherited lastname-surnames would require a tag, or some other mechanism, to tell bots this individual should not be treated as if they had a European style inherited surname. There should have been a template with a name like {{NoDefaultSortOk}} or {{NoListasOk}}. This should have been the default. And, I suggest, even though it is rather late in the day, it would make sense to establish a convention that the listasbot, the defaultsortbot, AWB, Kingbotk, and all other bots leave alone artcles with a tag that tells them the article is not someone with a European style name.
Candidly, Geo Swan (talk) 17:00, 28 May 2009 (UTC)[reply]
WP:TLDR. A long rant, examples of human editors screwing up, and assertions without evidence will not help your case. If you want something done, you'll have to provide a clear, concise suggestion of a positive course of action. Random assertions about the population of China mean absolutely nothing; more useful would be the number of articles on the English Wikipedia with "European" versus "non-European" names. Anomie 19:37, 28 May 2009 (UTC)[reply]
It's true that it's a problem. I hope that we can recruit people knowledgeable in Arabic (and Persian) conventions to confirm the sorting of these names. – Quadell (talk) 13:04, 27 May 2009 (UTC)[reply]

Category:Open Wikipedia bot requests for approval

There sure seem to be a lot more pages in Category:Open Wikipedia bot requests for approval than are listed as still open on this page. --Pascal666 05:44, 29 May 2009 (UTC)[reply]

I went through and removed those that had been approved/denied, but someone else needs to have a look see which if any have slipped through the net. - Jarry1250 (t, c) 06:38, 29 May 2009 (UTC)[reply]
There are only 15 there at the moment, and they all appear to still be open; nothing looks to have slipped through the cracks. Useight (talk) 03:38, 3 June 2009 (UTC)[reply]

I added "Estimated number of pages affected" to Wikipedia:Bots/Requests for approval/InputInit. It's something I've been meaning to do for ages. Let me know if there are any issues. --MZMcBride (talk) 22:29, 7 June 2009 (UTC)[reply]

What should be entered for any task that's intended to be an ongoing task? For example, User:WebCiteBOT, User:ClueBot, and so on. Anomie 23:47, 7 June 2009 (UTC)[reply]
n/a? Or perhaps 1 page per day or whatever. Obviously this won't apply to all bots, but it is something that comes frequently and the question is rarely asked enough. If it doesn't apply, bot ops can skip over it or write "n/a". Though I'd imagine even bots like ClueBot have an estimate of how many pages they edit per day. --MZMcBride (talk) 01:00, 8 June 2009 (UTC)[reply]
I've gone ahead and clarified that in the comment. Although really estimating for some bots would require a good deal of study to determine beforehand. Anomie 02:31, 8 June 2009 (UTC)[reply]
Looks good to me. :-) --MZMcBride (talk) 02:47, 8 June 2009 (UTC)[reply]

Lightbot

With the completion of the date delinking arbitration case, I was just wondering what next steps, if any, needed to be taken by BAG. Lightbot (talk · contribs) hasn't been banned directly, but that's the effective result of the remedies related to Lightmouse (talk · contribs). Just wondering if it's appropriate for BAG to revoke any task approvals for Lightbot that still exist, and ask a 'crat to remove the account's bot flag and block the account in accordance with the remedies? I also note that Lightbot isn't listed anywhere on the status page. Mlaffs (talk) 15:43, 15 June 2009 (UTC)[reply]

Yes, the 'crats should remove the bot flag and an admin should block the account, as Lightmouse is indefinitely prohibited from using any sort of automation. As far as I'm concerned, Lightmouse must (1) successfully appeal to ArbCom to have that prohibition revoked, and then (2) request (re-)approval before running any sort of bot in the future. Anomie 02:01, 16 June 2009 (UTC)[reply]
User:fl blocked it, but a crat still needs to deflag. – Quadell (talk) 12:51, 16 June 2009 (UTC)[reply]
Requested at WP:BN#Lightbot. Anomie 14:12, 16 June 2009 (UTC)[reply]
Done by Dweller Anomie 00:24, 17 June 2009 (UTC)[reply]

Why do we have a bot creating redirects that are handled by case insensitivity in the search box?

Resolved
 – Supplementary BRFA filed by operator to ensure community consensus exists for these redirects. –xenotalk 13:50, 25 June 2009 (UTC)[reply]
BOTijo (BRFA · contribs · actions log · block log · flag log · user rights)

See Wikipedia:Bots/Requests for approval/BOTijo 2. This may have been approved before case-insensitivity in the search field was implemented and the task may now need to be revoked - it seems to be creating many unnecessary redirects. –xenotalk 21:12, 24 June 2009 (UTC)[reply]

I've temporarily blocked the bot, I think this task should be stopped while this is looked at. –xenotalk 21:27, 24 June 2009 (UTC)[reply]
Hi Xeno. Please, read this (Other capitalisations, to ensure that "Go" to a mixed-capitalisation article title is case-insensitive). Regards. Emijrp (talk) 22:25, 24 June 2009 (UTC)[reply]
I've done a little experimenting, and it is still the case that an article with some words capitalized and some not will not be found by the Go button. Anomie 23:26, 24 June 2009 (UTC)[reply]
Can you give an example? All garblings of an initial-cap title like Francis Ford Coppola are supposed to be handled correctly by Mediawiki. Like fRancis fOrd cOppola, which works for me with the Go button, though it's a red link when bracketed in running text. This is explained in the link given above by Emijrp. EdJohnston (talk) 23:37, 24 June 2009 (UTC)[reply]
Yes, a title with all words capped will work. But try something like "barack obama speech to joint session of congress, 2009" which has a mix of capped and uncapped words (as stated pretty much everywhere discussing this issue); it won't bring you to Barack Obama speech to joint session of Congress, 2009. Anomie 00:51, 25 June 2009 (UTC)[reply]
Shouldn't this be solved by a fix to the software rather than having a bot create redirects non-stop? However, since this is a new facet that I was not previously aware of, I don't object to the bot resuming operations in the meanwhile. –xenotalk 00:58, 25 June 2009 (UTC)[reply]
The search box is not the only navigation tool. Many of us just type the page name we want directly into the URL. Also, redirects are usefully for linking. One can debate the merits of this task, but basing it on the search box alone is shortsighted. -- JLaTondre (talk) 23:46, 24 June 2009 (UTC)[reply]
  • Based on the new information, all I would like is an active BAG member to rubberstamp the BRFA. The closing statement didn't really give me any confidence in the community's approval for the task. –xenotalk 00:58, 25 June 2009 (UTC)[reply]
  • We have the search box in its present form, with its present limitations. We need a way of working around them, and this bot will do it. When we have a better search function, we can reconsider whether it is still necessary. Waiting for a fix in the software can take years. DGG (talk) 03:41, 25 June 2009 (UTC)[reply]
  • The software does not handle these automatically. It is quite often that I arrive at a nonexistent page because of miscapitalization. For example, as JLaTondre says, one can simply edit the URL bar to go to an article (or construct such a URL by software), and this does not work when there are capitalization issues. Also, redlinks are case sensitive: Bell Jar really ought to be a redirect to Bell jar or The Bell Jar. — Carl (CBM · talk) 03:58, 25 June 2009 (UTC)[reply]
    • Hmmm, I won't say it's anti-wiki to create these redirects automatically, but it can lead to a lot of noise. Natural creation of redirects is usually better because there's a demonstrable time when someone tried to navigate somewhere and failed. If, after eight years, nobody has bothered to create a redirect at "Bell Jar," I'm not sure I see a compelling reason to do so now with a bot. --MZMcBride (talk) 04:37, 25 June 2009 (UTC)[reply]
      • I create these when I run into them, usually. I would prefer if there was a bot doing it so I wouldn't notice them at all. But this bot request was even more specific (read: conservative) than that; it only covers things like Urban Adult Contemporary Airplay panel where it is truly painful to guess the right capitalization. — Carl (CBM · talk) 04:50, 25 June 2009 (UTC)[reply]
      • @CBM, yes, I see now that I was working from an erroneous misunderstanding of the case-insensitivity function of the search box. Nevertheless, I'm of the same mind as MZMcBride. While I see now why these redirects are necessary due to the WP:MIXEDCAPS issue, I don't think many of them will ever be used. Especially the "barack obama speech ..., 2009" mentioned above. Doubt someone really knows the full name to type into the search box for that =) –xenotalk 12:52, 25 June 2009 (UTC)[reply]
  • My 2¢: considering that the BRFA was nearly two years ago and generated very little attention at the time, it wouldn't hurt to create a new one for the task in order to properly judge community consensus. --ThaddeusB (talk) 05:15, 25 June 2009 (UTC)[reply]
See Wikipedia:Bots/Requests for approval/BOTijo 2 (again). Emijrp (talk) 13:40, 25 June 2009 (UTC)[reply]

Anybot's latest approval

The anybot owner says it didn't change the articles this time with the same bot but with "a different bot."[1] I don't see the bot approval discussion in Martin's edit history.[2]

Please explain the basics of this to me in relation to this group. It seems each bot requires approval by the board on the project page, then flagging by a bureaucrat to operate. And that this must be done for each new task. If this is "a different bot" shouldn't it have a request for approval?

There are a lot of subpages for this project. This post is about "requests for approval," so I think this is the correct place to ask this. Was this bot approved? Did it require approval? Or was I mistaken in thinking that because the bot was blocked the issue could safely be discussed without additional problems arising? --69.226.103.13 (talk) 07:23, 25 June 2009 (UTC)[reply]

Request for deflagging and blocking of Anybot

In accordance with Wikipedia:Bot policy#Appeals and reexamination_of_approvals, this is a formal request for Anybot to be deflagged and indefinitely blocked.

Anybot has now had four major runs. The first run, in February, introduced many major errors, by admission of Martin, the bot operator.[3] The second run, in March and April, fixed some of these errors; but it didn't even come close to making these articles acceptable. From April on, Martin was being asked to address problems introduced by his bot, and did not do so. For example, on 6 March Rkitko pointed out that Anybot had wrongly described thousands of cyanobacteria as algae[4], and raised the matter again on 21 April[5], but as of today, 26 June, Rkitko hasn't received a reply[6] and these articles have still haven't been fixed.[7]

Anybot ran for a third time in May and June, and continued to introduce errors. It also exhibited unacceptable behaviours such as edit warring.[8][9] Martin has stated that he did not run the bot at this time, and that whoever did run it was not authorised to do so; apparently anyone could run the bot by visiting a certain webpage; he did not bother to secure the page because he figured no-one knew of its existence—security through obscurity.[10][11]

The extent of the problem did not become clear until the last couple of weeks, when 69.226.103.13, who appears to have expertise in this area, spoke out strongly on the matter at WT:PLANTS. There was a long discussion, during which it became clear that there were so many wrong articles, with so many errors, of some many different types, that the only way they could be fixed is they were individually manually repaired by a phycologist. This would take thousands, perhaps tens of thousands, of hours; it would probably be quicker to delete them and write them all from scratch. Therefore I sent all 4000 articles to AfD; consensus seems to be emerging there that they will need to be deleted.[12]

One result of the AfD discussion was that it finally prompted Martin to respond. Having discovered that the bot had been run without his authorisation, he blocked it. He then began working on a bot that would fix the errors. Once this bot was ready, he announced his intention of running it. A number of people objected to the idea that Anybot could be trusted to fix these errors.[13][14][15] But despite these objections, and calls for the bot to be deflagged,[16][17][18][19] Martin unblocked the bot and set it going, apparently without a test run, and without notifying or seeking approval from the BAG.

This fourth run put a great many articles into a novel state, including introducing new errors, such as classifying diatoms as plants.[20] These were all new edits, not reverts; but disturbingly, every edit was marked as minor, and given the misleading edit summary "Restore article to last good version."[21] The bot also edited at least one article that it had never edited before,[22] despite Martin's assurance that it had only edited articles created by Anybot and not since edited by a human.[23] I have now reblocked the bot.

In summary, this bot has been a complete disaster from start to finish. Martin may have the best of intentions but he has presided over a monumental screwup and his bot cannot be trusted at any level. I am seeking to have Anybot deflagged and indefinitely blocked on the grounds that

  • it introduces major errors of facts on a massive scale, every time it is run;
  • it has exhibited unacceptable behaviours such as edit warring and the use of misleading edit summaries;
  • the bot owner failure to secure the bot account;
  • the bot owner failed to address and fix errors in a timely manner;
  • the bot owner has unblocked and run the bot in the face of community opposition to him doing so.

Hesperian 03:05, 26 June 2009 (UTC)[reply]

anybot (talk · contribs · count) 's BRFA approval page Wikipedia:Bots/Requests_for_approval/anybot -- Tinu Cherian - 04:50, 26 June 2009 (UTC)[reply]
Comment: Anybot is currently indefinitely blocked. – Quadell (talk) 13:11, 26 June 2009 (UTC)[reply]
See Wikipedia:Articles_for_deletion/Anybot's_algae_articles#Solution. Martin (Smith609 – Talk) 18:54, 26 June 2009 (UTC)[reply]
Sounds like an appropriate solution. Unless any other BAG members object, I'll mark Wikipedia:Bots/Requests for approval/anybot as revoked in a few hours.
As for the deleting, whoever tags the pages for deletion (or runs an adminbot to just delete them) should have that task approved in the normal way; it's not so urgent that WP:IAR is needed, IMO. If no one beats me to it, I may write a quick script to generate the list of pages needing deletion, at which point any admin approved for mass-deletion (e.g. with AWB) could handle it. Anomie 22:24, 26 June 2009 (UTC)[reply]
It is done. Anomie 03:04, 27 June 2009 (UTC)[reply]
Sorry Anomie, but if the community reaches the consensus that they should all be deleted, then they all get deleted. You don't get to end-run around the outcome of the AfD. If you read the AfD carefully, you will find reasons given why it is insufficient to "delete all articles which have only been edited by Anybot (and maintenance bots such as User:Addbot)". In short, the articles must be assumed to have been error-ridden at the time of creation, and the fact that some have been edited since does not imply that they have corrected or verified; more than likely the subsequent edits were merely cosmetic, since there are very few editors here with the expertise to contribute content or corrections on this topic. Hesperian 04:47, 27 June 2009 (UTC)[reply]
I don't give a crap about AfD, and the unapproving of anybot is exactly what you asked for. What's your point? Anomie 04:49, 27 June 2009 (UTC)[reply]
Didn't you just scrape a list of articles created by Anybot and not since edited by a human? That's what Martin is proposing in the linked section. And what I read from you above is "Sounds like an appropriate solution.... I may write a quick script to generate the list of pages needing deletion.... It is done." To me, that sounds like a declaration that you intend to implement Martin's solution. My point is that the AfD process decides what to do with these articles, not you. Hesperian 05:11, 27 June 2009 (UTC)[reply]
Actually, the list isn't completed yet. Whether it actually gets used for deleting the pages or not, I don't really care; I'll just post it (actually, it'll be at least 3 lists: created by anybot and edited only by bots, created by anybot and edited by non-bots, and just edited by anybot) somewhere. I actually forgot about that when I said "It is done", I was referring to the revoking of the bot's approval. Anomie 05:15, 27 June 2009 (UTC)[reply]
I see. At present the AfD is based upon the list at User:Anybot/AfD, which was originally a list of articles created by anybot, but has since been refined by the removal of some articles that have been fixed by a human. Therefore a "created by anybot" list would be redundant. But as you say, a "created by anybot and edited only by bots" list would enable the implementation of Martin's proposal, should it be agreed to. Hesperian 05:29, 27 June 2009 (UTC)[reply]
Here are the lists:
Do with them what you will. Anomie 05:49, 27 June 2009 (UTC)[reply]
I have removed Anybot's flag. bibliomaniac15 05:04, 27 June 2009 (UTC)[reply]
Thank you. --69.226.103.13 (talk) 06:29, 27 June 2009 (UTC)[reply]

You know, it would be nice if these articles were ever examined in a way that showed fewer errors instead of more in number and kind. The list with articles that anybot touched but did not create contains a whole new mess of errors, each unique, and each will have to be checked and fixed by a human.

Also, the bot's synonymies are wrong, so probably 100% of its redirects should also be deleted if they can't be 100% checked, although user Hesperian is planning to deal with that.

I hope BAG makes certain that bot created articles in the future are coded properly to NOT overwrite existing articles.[24][25]

It's hard to understand how the bot was ever allowed to continue the first time it was noticed it was doing this. Obscure algae require expertise programmers may not have, but code that allows a bot to overwrite existing, unrelated text, is a major and inexcusable error.

This group tends to ignore comments made by IPs-you don't respond to my posts. But, if you, as a group, did not ignore IPs, someone might have caught and stopped this mess long before it reached this level. The IP 213.214.136.54 edited over a thousand articles, correcting the most egregious errors, and all of his/her edits and hard work are slated to be deleted.

IPs contribute a lot of excellence to wikipedia. I can't stop you from ignoring my every post, and setting an example to bot operators that this is how to act (as Martin acted), but the wikipedia community has decided over and over to allow anonymous IPs to edit.

If this group does not respect the community consensus, it's no wonder that it allows the creation of messes that put wikipedia in disrepute.

A group that can make this much work for other writers of the encyclopedia should be a part of the community, not a non-responsive law alone.

That's just my opinion on the matter. --69.226.103.13 (talk) 06:29, 27 June 2009 (UTC)[reply]

With all do respect, your comment is way off base. No one "ignored IP input", as there was none to consider in the original BRFA. BAG members aren't some kind of super geniuses that can automatically envision every possible problem about every possible subject. The bot was correctly approved based on the feedback that was received during the approval process (which was pretty extensive, by the way). Trying to blame the bad articles on BAG is ridiculous and completely not helpful to solving the actual problem. --ThaddeusB (talk) 15:43, 27 June 2009 (UTC)[reply]
I have two posts on pages by this group that have been completely ignored. One just above this thread.
It doesn't require a super genius in programming to understand that a bot that is creating articles should not be coded to overwrite entire other articles without being specifically coded for that feature.
You're the final say in authorizing a bot, before it's flagged, but you don't monitor problems with the bot; there's nothing way off base about my comments. --69.226.103.13 (talk) 17:22, 27 June 2009 (UTC)[reply]
The BAG (of which I am not a member, BTW) can only go by what it sees and no objections or evidence of problems were raised during the BRFA. It is unfortunate that the bot made a large number of errors and that the operator failed to secure to code, which allowed a vandal to abuse it. However, both of those problems ultimately fall on the bot owner, bot BAG. BAG can't predict the future and expecting them to do so is completely ridiculous. --ThaddeusB (talk) 19:28, 27 June 2009 (UTC)[reply]
If there were consequences to bot operators who ignored writer-alerts about problems with their articles, then this problem area could be reduced. Note, someone said Martin corrected errors. He did not; he ignored error reports for months, and his early reports of correcting errors are wrong; and, if he had examined the results, he could have seen that he sometimes made the problem worse and seldom fixed it.
Now, it seems once the BAG approves your bot you can do what you want: restart your bot when it's been blocked for a reason, run the bot to do unapproved tasks against the community consensus, and create new bots and operate them without consequences once you have a flagged bot. All of this is a problem with BAG: there are no consequences to bot owners operating their bots in unapproved manners, thus giving tacit approval to bot owners to do as they please. Until BAG requires bot owners to monitor their bots, secure their codes (a simple and basic requirement), and be responsive, in a timely fashion, to bug reports, this problem will plague wikipedia and its articles, creating future messes. --69.226.103.13 (talk) 20:09, 27 June 2009 (UTC)[reply]
His bot got blocked. Then, someone assumed good faith and unblocked it. And how on Earth is the BAG meant to force bot owners to monitor their bots anyway? - Jarry1250 [ humourousdiscuss ] 20:18, 27 June 2009 (UTC)[reply]

Arbitrary break

Replies below are chronologically before many of the replies above. This break splits up comments and replies, which is unfortunate. - Jarry1250 [ humourousdiscuss ] 20:18, 27 June 2009 (UTC) [reply]

What you have said of substance has been lost in long rants about how horrible the bot is and how horrible we are, which is why I haven't replied to you yet. I don't see that 213.214.136.54 has commented here anywhere. Anomie 13:56, 27 June 2009 (UTC)[reply]
I have not posted long rants about how horrible you are. I posted a series of questions above, while trying to understand is there some way this problem can be prevented in the future. I posted another question elsewhere. You've ignored my questions. If you can't or won't read a systematic evaluation of what is wrong with a bot in an attempt to find a way to fix the problem with programming, then this group is being irresponsible. That's nothing to brag a link about.
I posted lengthy discussions about the problems with the bot because I wanted to be sure I identified all types of errors, initially, when it was thought the errors could be corrected. When I realized they couldn't I posted a sufficient number of examples of the types of errors to make it clear that there was not a set of systematic errors that could be corrected with another bot.
I continued to keep searching for this type of error, something that could save a large group of the articles.
What I've said in my rants is germane to the discussion as long as writers continue to try to save a large quantity of the articles.
No, that's not why you or other members of this board haven't replied to my posts here, such as my questions above: it's because you ignore the input of IPs. This is typical wikipedia from my experience, including, now, this one: trying to get something done about over 4000 really bad articles. Bad in their entirety, bad they should have never been here articles.
IPs are part of the wikipedia community. My rants (thanks for the personally insulting sling about my work) were necessary to describe the massive extent of the errors created by this bot. They were often in response to other writers who were trying to save articles. When other writers had suggestions, rather than ignoring them I attempted to find ways of implementing them by reviewing more articles. I posted the results of these reviews, which you call rants. I guess I should have been thankful to be ignored by members of this group, it seems it was nicer than the personally insulting responses based on ignorance of the work I've done trying to save some of the articles by finding a large group of bot-editable ones.
If I'm missing a compliment I should be giving, please let me know.
Other IPS have posted to Martin's bot alerts board about article problems and been ignored.
I hate to keep having to say this, but, please, stay away from insulting me and focus on the substantive issue.
If my paragraph in the thread above is too long for members of this board to read, then command that folks write messages of only 1 or 2 simple sentences. --69.226.103.13 (talk) 17:22, 27 June 2009 (UTC)[reply]
Again, no one ignore you because you are an IP. This is a completely unfounded accusation. --ThaddeusB (talk) 19:28, 27 June 2009 (UTC)[reply]
We have focused on the substantive issues. See above. Anomie 17:24, 27 June 2009 (UTC)[reply]
The major substantive issue dealing with programming is preventing a mess of this nature from ever putting Wikipedia in disrepute of this level again. That issue has not been touched by this group. --69.226.103.13 (talk) 17:33, 27 June 2009 (UTC)[reply]
So what do you propose we do? The bot was given 2 trials, and all the reported errors from the trials were supposedly fixed. Ultimately, the responsibility for the bot rests on the operator. Your post above was probably ignored because you gave almost no context for your comments. Mr.Z-man 17:50, 27 June 2009 (UTC)[reply]
I think the question of running an unapproved bot was enough that clicking on the links for context was in order. This isn't the first time Martin ran a bot without approval.
There are always a million more ways to work as a group to contribute something positive than there are ways to simple flame war.
A short term solution to part of the problem: Someone mentioned a {{noindex}} tag would prevent the articles from appearing in google search results. While the articles are still being discussed, even for only one more day, can someone who runs bots immediately tag all of the article's on this page with noindex? This does not need to be done to articles not created by anybot, just the ones it created (its deletion edits are among the ones it edited but did not create).
This could serve a couple of immediate goals. The edit summary of the bot tagging could link to the AFD list (User:Anybot/AfD) in its explanation for why a no-index tag was added. Anybot's user AfD list now links to the AfD. It could also alert anyone watch-listing any of the articles about the AfD and the need to remove any of their articles from the list.
--69.226.103.13 (talk) 18:08, 27 June 2009 (UTC)[reply]
The BAG doesn't have a magical ability to prevent someone form running an unapproved bot. There is no technical limitation that the BAG enables to allow a bot to run. --ThaddeusB (talk) 19:28, 27 June 2009 (UTC)[reply]
Yet there are no consequences to running an unapproved bot. A post about an unapproved bot was ignored. It's not even interesting that someone runs an unapproved bot. --69.226.103.13 (talk) 20:09, 27 June 2009 (UTC)[reply]
Your assertion that BAG members would ignore an unapproved bot is absurd. The bot that caused these problems was an approved bot. You fundamentally misunderstand the difference between changing an aspect of a bot's code and a completely new task. Of course someone doesn't have to seek approval every time they modify their code. Bots are approved for a task and it is perfectly permissible to change one's code to better do the same task without seeking new approval.
Also, many users have been banned from Wikipedia or otherwise disciplined for running unapproved bot, so there certainly is not "no consequence" as you assert. --ThaddeusB (talk) 21:00, 27 June 2009 (UTC)[reply]
It was me who suggested that. I actually thought that I'd proposed noindexing the articles earlier this week - but looking back at my contribs, this seems not to be the case (perhaps I got edit conflicted? dunno). Unless this is prohibited by policy somewhere, this would be a Very Good Idea, IMO. 69.226. is correct in saying that these broken articles don't exactly make WP look good... --Kurt Shaped Box (talk) 18:18, 27 June 2009 (UTC)[reply]
I wish you had. You may have, and maybe I missed it. Let's get it done now if possible, though. --69.226.103.13 (talk) 18:20, 27 June 2009 (UTC)[reply]

On the question of what can you do? How about implement the most basic type of rules that programmers use when paid to write code? A simple start is demanding algorithms, maybe only from new programmers. Some programmers have many bots and few errors.

Any competent programmer can read another's algorithm and see they've missed the most basic things like initializing variables (probably why Martin's bot added bad lines of text and created bad taxoboxes, if the information had been gathered from a prior genus, but the next genus didn't mention spore types or its taxonomy, it just used leftover information), no protection against deleting an entire existing article. This is not genius level programming.

Then get specific approval from interested editors in their arena for all bots creating articles-don't just wait to see if anyone objects here, get positive approval. After a bot's initial run creating articles, post the list of its articles on wikiproject plants or somewhere and ask writers to check off on all of the articles.

--69.226.103.13 (talk) 18:20, 27 June 2009 (UTC)[reply]

I think the anybot edits show allowing a bot to create articles has potential to create an even bigger mess than this one. This appears to be because of a lack of basic coding requirements and fundamental safeguards from the BAG prior to and once a bot is given approval.

Some writers are now battling me verbally to fight this realization. It won't change the situation as it is: bots on wikipedia are not controlled.

As someone else pointed out, if a meat editor had created these articles they would have been permanently blocked as a vandal once they were told of the problem and failed to stop. This block, in the case of anybot, would have occurred after the test run, not after 4000 articles.

I don't think bot owners want to hear this, and I've said my piece, and taken enough insults in exchange. This is a familiar wikipedia response to an editor, particularly an IP or newly registered user, pointing out a problem with wikipedia. This is how wikipedia winds up again with egg on its face in the news: wikipedia editors don't want to hear what's wrong. That's why this mess wasn't cleaned up in February: established editors refused to listen, a phycology editor, and the bot owner, and once a bot is approved its owner is allowed to do what he wants. --69.226.103.13 (talk) 20:21, 27 June 2009 (UTC)[reply]