Wikipedia:Bots/Requests for approval: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
→‎[[User:VoABot II]]: -->expired requests
(4 intermediate revisions by the same user not shown)
Line 501: Line 501:




==[[User:SmackBot|SmackBot]] task approval VI==
Task: To perform other formatting fixes to ISBNs. (SB already removes ":" after the "ISBN".) The additional functionality for example
*Delinks [[ISBN]] where it's followed by an ISBN
*Debolds/deitlicises "ISBN" where it's followed by an ISBN
*Replaces blanks in ISBNs with "-".
Method: Automatically using AWB or otherwise.<br>
Speed: 1-4 per minute.<br>
Number: c. 100 per month.<br>
Frequency: Approx every data dump <br>
Duration: About an hour<br>
Testing: A one-off cleanup has been done<br>
''[[User:Rich Farmbrough|Rich]] [[User talk:Rich Farmbrough|Farmbrough]]'' 06:30 [[26 July]] [[2006]] (GMT).
:One week (up to a month if needed) trial approved, restrict the trial of this new fucntion to 100 edits, and post results when trial complete. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 16:31, 29 July 2006 (UTC)
::Cheers. ''[[User:Rich Farmbrough|Rich]] [[User talk:Rich Farmbrough|Farmbrough]]'' 13:37 [[3 August]] [[2006]] (GMT).
::OK, got a database dump yesterday, here are a hundred sample edits. [http://en.wikipedia.org/w/index.php?title=Special:Contributions&limit=100&target=SmackBot] ''[[User:Rich Farmbrough|Rich]] [[User talk:Rich Farmbrough|Farmbrough]]'' 00:23 [[15 August]] [[2006]] (GMT).
===Awaiting approvals group response===


----


==[[User:BOTepho]]==
==[[User:BOTepho]]==
Line 611: Line 594:


=Approved, needs to be flagged=
=Approved, needs to be flagged=

----
----


=Approved, flagged=
=Approved, flagged=
<div class="boilerplate metadata" style="background-color: #dedaca; margin: 2em 0 0 0; padding: 0 10px 0 10px; border: 1px solid #AAAAAA;">
:''The following discussion is an archived debate. <span style="color:red">'''Please do not modify it.'''</span> Subsequent comments should be made in a new section.'' <!-- from Template:Debate top-->

==[[User:SmackBot|SmackBot]] task approval VI==
Task: To perform other formatting fixes to ISBNs. (SB already removes ":" after the "ISBN".) The additional functionality for example
*Delinks [[ISBN]] where it's followed by an ISBN
*Debolds/deitlicises "ISBN" where it's followed by an ISBN
*Replaces blanks in ISBNs with "-".
Method: Automatically using AWB or otherwise.<br>
Speed: 1-4 per minute.<br>
Number: c. 100 per month.<br>
Frequency: Approx every data dump <br>
Duration: About an hour<br>
Testing: A one-off cleanup has been done<br>
''[[User:Rich Farmbrough|Rich]] [[User talk:Rich Farmbrough|Farmbrough]]'' 06:30 [[26 July]] [[2006]] (GMT).
:One week (up to a month if needed) trial approved, restrict the trial of this new fucntion to 100 edits, and post results when trial complete. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 16:31, 29 July 2006 (UTC)
::Cheers. ''[[User:Rich Farmbrough|Rich]] [[User talk:Rich Farmbrough|Farmbrough]]'' 13:37 [[3 August]] [[2006]] (GMT).
::OK, got a database dump yesterday, here are a hundred sample edits. [http://en.wikipedia.org/w/index.php?title=Special:Contributions&limit=100&target=SmackBot] ''[[User:Rich Farmbrough|Rich]] [[User talk:Rich Farmbrough|Farmbrough]]'' 00:23 [[15 August]] [[2006]] (GMT).
* All difs look fine, keep an eye out for any side afects that AWB "general fixes" may have, watch the AWB page if you will use this feature. This bot should be flagged. Until flag is set, limit edits to !>3/min. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 15:01, 19 August 2006 (UTC)
**Just noticed you were already flagged, full bot operations are a go. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 15:04, 19 August 2006 (UTC)
:''The above discussion is preserved as an archive of the debate. <span style="color:red">'''Please do not modify it.'''</span> Subsequent comments should be made in a new section.''<!-- from Template:Debate bottom --></div>

----
----
<div class="boilerplate metadata" style="background-color: #dedaca; margin: 2em 0 0 0; padding: 0 10px 0 10px; border: 1px solid #AAAAAA;">
<div class="boilerplate metadata" style="background-color: #dedaca; margin: 2em 0 0 0; padding: 0 10px 0 10px; border: 1px solid #AAAAAA;">
Line 886: Line 890:
<!--Place newest entries here at the top-->
<!--Place newest entries here at the top-->
----
----
<div class="boilerplate metadata" style="background-color: #dedaca; margin: 2em 0 0 0; padding: 0 10px 0 10px; border: 1px solid #AAAAAA;">
:''The following discussion is an archived debate. <span style="color:red">'''Please do not modify it.'''</span> Subsequent comments should be made in a new section.'' <!-- from Template:Debate top-->

==[[User:VoABot II]]==
This bot is basically part of what VoABot did, but I decided to split the functions up. This bot watches certain pages and reverts edits using regexps denoted as blacklisted. It can check for added content, summaries, and logged-out IP ranges, depending on what is needed to stop recurring AOL/Shared IP vandals. Its been running for a while as part of VoABot[http://en.wikipedia.org/w/index.php?title=Special:Contributions&limit=100&target=VoABot&namespace=0], and I'd just like to have this second account flagged, since it can sometimes make a lot of edits in a fairly short time period (though nothing like Rambot). Here[http://en.wikipedia.org/wiki/WP:PP#Protected_against_spambots_or_banned_users] are the pages that it watches (it updates the list hourly).'''[[User talk:Voice of All|<font color="blue">Voice</font><font color="darkblue">-of-</font><font color="black">All</font>]]''' 03:23, 28 July 2006 (UTC)
:While I support the idea that these functions should be from separate accounts, I'm missing where the reversion function of VoABot was approved, the [[Wikipedia:Bots/Requests_for_approvals/Archive2#VoABot|original]] request does not mention this "feature". What are the parameters and other tests used to prevent bad reverts here? The page for this bot also states that it has already been approved here??? — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 04:08, 28 July 2006 (UTC)

:Also just saw edits like [http://en.wikipedia.org/w/index.php?title=Qipao&action=history these], that don't look like '''obvious vandalism''', with no discussion going on on the talk page. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 04:19, 28 July 2006 (UTC)
::VoABot's page says that its functions are approved, but not VoABot II. VoABot IIs functions were tested as part of VoABot for a while. Also, it does not try to find general vandalism, as TB2 and the like already due. It goes after edits typical to specific banned users or AOL attacks. The edits you were looking at were by a revolving IP adding the same consensus-rejected spam links for weeks. That and Gibriltarian-like stuff is what this reveerts. It uses the standard rollback method (revert to last non-X contrib, check if X is actually the last contrib) and whitelists admins.'''[[User talk:Voice of All|<font color="blue">Voice</font><font color="darkblue">-of-</font><font color="black">All</font>]]''' 05:05, 28 July 2006 (UTC)
::[http://en.wikipedia.org/w/index.php?title=User%3AVoABot_II&diff=66297559&oldid=66287469 Removed note] (assuming now it was a copy and paste of old bot page.) Wouldn't sprotecting the page be more effictive then revert warring with an anon though? — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 05:15, 28 July 2006 (UTC)
:::Usually they just give up after a while. Also, they may just make throwaway accounts to get through, which does little if the edits are reverted promptly. Additionally, the AOL/shared IP RC patrol has stopped several trolling streaks to many randomn pages, something sprotection is useless against. Also, the note you removed was a legend, all the checked features where approved (VoABot II had no checked feautres).'''[[User talk:Voice of All|<font color="blue">Voice</font><font color="darkblue">-of-</font><font color="black">All</font>]]''' 05:22, 28 July 2006 (UTC)
*Voice of All is careful and intelligent. I think he would make a great bot. —[[User:Centrx|Centrx]]→[[User talk:Centrx|''talk'']]&nbsp;&bull; 20:01, 29 July 2006 (UTC)
*Edits such as [http://en.wikipedia.org/w/index.php?title=Cat&diff=prev&oldid=66462599 this one] seem to violate [[WP:BITE]] and [[WP:AGF]], and are being made without so much as a talk message, with a edit summary that the edit is RESTRICTED. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 22:35, 29 July 2006 (UTC)
**I've changed the regexps "from previous vandal streaks" so it wont revert that edit again (or any edit with 7 or so periods in a row). I've modified the edit summaries to include "vandalism" and "spam links". Also, what should the edit summary be changed to? Should it notify all non-AOL shared IP users (as notifying AOL users is pointless as the adress changes)?'''[[User talk:Voice of All|<font color="blue">Voice</font><font color="darkblue">-of-</font><font color="black">All</font>]]''' 22:59, 29 July 2006 (UTC)
***You may want to check with the Tawkerbot team, they have a notification sytem built in to their reverts, as for the edit summary, telling someone that their editing is restricted goes against the "...that anyone can edit", without any forwarning (like {{t1|sprotect}} does. What some other have dne is create a subpage on your bot, and have the edit summary link to it, where the page explains to the user why they were reverted (generically of course) and how they should avoid getting reverted again (assuming good faith, and not requiring that they get an account). — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 23:14, 29 July 2006 (UTC)
****Subpages, good idea. I'd work on those now. I'd like to get the RC patrol for this bot on again asap, as AOL vandal streaks can really come hard sometimes.'''[[User talk:Voice of All|<font color="blue">Voice</font><font color="darkblue">-of-</font><font color="black">All</font>]]''' 23:26, 29 July 2006 (UTC)
****OK, I started the subpage, and the edit summaries now link there.'''[[User talk:Voice of All|<font color="blue">Voice</font><font color="darkblue">-of-</font><font color="black">All</font>]]''' 00:48, 30 July 2006 (UTC)

*What little I've seen of this bot, it's done a very good job and stopping IP hopping aol vandalism.(Where a vandalbot hops all over the place on a subrange vandalising). I like the basic idea. Though xoasflux is probably correct about how socialable it is, its technical capabilities seem quite accurate. You should fix it up a little, and it will be a nice compliment to what the tawkerbots do. [[User:Kevin Breitenstein|Kevin_b_er]] 01:15, 30 July 2006 (UTC)

*One week trial approved, please maintain a record of any complaints (except those from '''obvious''' vandals), and link to them (if any) along with some difs here. Don't get discouraged if someone blocks you along the way, even the tawkerbots got blocked at the start of their runs. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup> 14:19, 30 July 2006 (UTC)
:Too many times I might add (well, Tawkerbot2 got the worst of it, Tawkerbot4 only got a user block methinks) -- [[User:Tawker|Tawker]] 05:15, 31 July 2006 (UTC)
===Awaiting Operator Response===
*Moved to expired {14:56, 19 August 2006 (UTC)}, no response from operator. — [[User:Xaosflux|<b><font color="#FF9933" face="monotype"><big>xaosflux</big></font></b>]] <sup>[[User_talk:Xaosflux|<font color="#00FF00">Talk</font>]]</sup>
{{subst:debate borrom}}
----





[[bg:Уикипедия:Бот/Заявки]]
[[bg:Уикипедия:Бот/Заявки]]

Revision as of 15:04, 19 August 2006

New to bots on Wikipedia? Read these primers!

To run a bot on the English Wikipedia, you must first get it approved. Follow the instructions below to add a request. If you are not familiar with programming consider asking someone else to run a bot for you.

 Instructions for bot operators

Current requests for approvals

A manually-initiated pywikipedia-based bot which would be used for template substitution and commenting out images. It will only run when it is given an image or template to run on by me. It will comment out images that are speedy deleted, deleted because of IFD, or deleted because the copyright or source is unknown after the image has been tagged as such for seven days, and will substitute templates listed at Wikipedia:Template substitution#Templates that should be substituted and template deletion debates that end in "substitue and delete". It will get the pages using the image from the "File links" section of the image page and the pages using the template from the "What links here" link. —Mets501 (talk) 15:16, 18 August 2006 (UTC)[reply]

Surely speedy deleted and ifd'd images should be removed altogether rather than commenting out? Martin 15:21, 18 August 2006 (UTC)[reply]
Whichever is more correct; I can have it remove the image altogether for IFDs and speedy deletes and comment it out for unsourced images or images of unknown copyright status. —Mets501 (talk) 15:23, 18 August 2006 (UTC)[reply]

Also, I'd like to also let this bot run on AWB or the pywikipedia framework for SI writing style adjustments (see International System of Units#SI writing style), and for other simple, uncontroversial changes (such as changing Maori to Māori, etc.) —Mets501 (talk) 14:14, 19 August 2006 (UTC)[reply]

A manually-operated script, in Java with WikiLink. For starters, it will generate a list (on a user subpage) of website stubs that are not in Category:Websites (or a subcategory). If approved and successful, it could eventually be extended to other categories. Should be a fairly simple script without room for damage, since it won't be making any mainspace edits. ~ Booya Bazooka 01:12, 15 August 2006 (UTC)[reply]

You need to explain this a bit more. What is the purpose in what you are doing, that is what value are you gaining with that information what do you intend to do with it. Also since Website stubs is a subcategory of Category:Websites isn't the product of this going to be an empty list? --pgk(talk) 06:57, 15 August 2006 (UTC)[reply]
The aim is to make sure that all of these website articles get categorized; having them all piled up in the stub category isn't very useful to someone browsing the categories. So, it would help greatly to just see a list of the articles to which I need to add categories. To answer your second question, the subcategory list will exclude Website Stubs. ~ Booya Bazooka 15:05, 15 August 2006 (UTC)[reply]
Ok as you say it should be fairly straightforward and only editing with it's own user name space. For implementation I'm not familiar with WikiLinks implementation, but you might like to consider query.php as a fairly simple and reliable way of extracting the category memberships. I assume this will run and generate the details and perform a single edit to refresh the page? How often are you anticipating to run this? Assuming you aren't intending to do 1000s of updates in order to achieve the result then go ahead with a trial. --pgk(talk) 17:07, 15 August 2006 (UTC)[reply]
I doubt I'll be running it any more than once a day, single edit. I am currently looking into query.php for the categorization. ~ Booya Bazooka 18:04, 15 August 2006 (UTC)[reply]

The bot uses the pywikipedia framework and is run manually. I'm planning to use it for multiple minor edits in relation to my organizational work on WP. First off, I'm looking to replace Cone (solid) links with Cone (geometry) and other updating tasks in relating to work on Cone (geometry), Right circular cone and Conic section.

I don't intend to use this bot a lot as I mainly plan on using the scripts on Wikibooks. The work here does not require the work of a bot, but it would save me a considerable amount of time. --Swift 05:49, 11 August 2006 (UTC)[reply]

If it is just a redirect-replace bot, I'm sure another bot could handle the job, but if you want to run it yourself, I don't see a problem with this. Are there any other jobs you're planning to do? Titoxd(?!?) 07:00, 15 August 2006 (UTC)[reply]
My rationale for running it myself is simple time saving — not having to request bot work which would slow down work and add load on others.
I have no other plans for the bot at the moment. I do, however, plan to continue editing along those lines, cleaning up articles and would use the bot for edits similar to those described. If I deceide to expand on the bot's tasks, I will of course request approval for those. --Swift 20:56, 15 August 2006 (UTC)[reply]
That didn't take long: I've already got the bot's second proposed task. If the Wikibooks template merge proposal goes through, I'd like to use this bot to update transclution links. --Swift 02:58, 18 August 2006 (UTC)[reply]

I have used this account to add templates to Dutch municipalities, and later to add articles about Dutch villages and towns. Now I want to create articles for Italian municipalities (see the example at User:Eubot/Moretta). Until now, the bot has run without a bot flag without any complaints. Because there are a lot of Italian municipalities to be added, I would like my bot to have a bot flag now. Eugène van der Pijll 21:13, 1 August 2006 (UTC)[reply]

Looks good, out of interest, where is the information from and what software do you use? Martin 21:16, 1 August 2006 (UTC)[reply]
The information ultimately comes from istat, the Italian statistical institute. I have not yet downloaded that data, however; these same articles have been added to the Dutch wikipedia recently, and I have asked the bot owner over there if he can provide all of the data in an easy-to-use format. I'll be using a Perl script, written completely by myself. Eugène van der Pijll 21:23, 1 August 2006 (UTC)[reply]
Would you mind providing a link to that institute website and tell us what license it uses for those datas? --WinHunter (talk) 02:19, 2 August 2006 (UTC)[reply]
I've been told the data on Italian municipalities has been uploaded to the Italian wikipedia: it:Wikipedia:Data/Comuni. I will probably use those files as source. The copyright page of Istat says: Dati ed analisi dell’Istituto nazionale di statistica possono essere scaricati, utilizzati e pubblicati a condizione di citarne la fonte. ("Data... can be used... on the condition of citing the source.")
This kind of factual/statistical data is ineligible for copyright in the United States, though not necessarily in other countries. Dragons flight 22:11, 2 August 2006 (UTC)[reply]
Seems good (as one would hope, given the name) on the basis of user-space drafts. Can you clarify under what circumstances the bot will be replacing existing articles, as opposed to adding new ones? And can you also give estimates of the number of articles to be created (or modified) on a per-region or per-province basis? Thanks. Alai 17:40, 4 August 2006 (UTC)[reply]
I don't know yet if I will replace articles automatically. For some regions in Italy, a lot of articles have been created containing nothing more than one sentence: "<<Town>> is a commune in <<Region>>, Italy." (e.g. Castelcivita). If there aren't too many variations, I can detect that and just overwrite those articles.
If the article is any longer than that, I won't do that. Instead, I'll just add it to a list, and decide later, interactively. I may decide to do that for all existing articles if I cannot detect the replaceable substubs with any accuracy.
There are over 8,000 municipalities in Italy; I'd guess no more than 2,000 have articles, half of which are completely replaceable substubs. So in the order of 7,000 articles in total. I don't have the exact numbers per region; the number of missing articles per province can be seen from templates like {{Province of Salerno}}. (That's actually a bad example, because all of those articles have already been created; but they are almost all substubs.) Eugène van der Pijll 19:05, 4 August 2006 (UTC)[reply]

I've finished writing the bot now, and I've done a small test run (15 articles) on the first links of {{Province of Turin}}. I think these articles are as good as I can make them, so I would like to have approval for this bot now. To answer one of the questions above: currently I'm not replacing any existing articles at all. Eugène van der Pijll 11:57, 12 August 2006 (UTC)[reply]

I have received a comment on User talk:Eubot/Moretta about moving a lot of data from the article to a (single-use) template. Would like feedback on that idea. Eugène van der Pijll 21:27, 13 August 2006 (UTC)[reply]
This bot seems to be running more then 15 articles, does this have approval of any kind? —— Eagle (ask me for help) 17:14, 17 August 2006 (UTC)[reply]
[1]. Eugène van der Pijll 17:19, 17 August 2006 (UTC)[reply]
If I am not mistaken that link shows your request for to run the bot; has it been actually been approved? - If not I strong suggest it ceases activity immediately, as according to this is has done over 1700 edits. - GIen 17:37, 17 August 2006 (UTC)[reply]
Automatic approval if there are no comments. (Old rules). Eugène van der Pijll 17:40, 17 August 2006 (UTC)[reply]

Actually, old rules don't apply but you are cleared for a trial and limited. Things look good though, run it nice and slow, its not something we want botflagged -- Tawker 17:53, 17 August 2006 (UTC)[reply]

Old rules do apply, as my bot is that old. But thanks nevertheless, I will finish the province I'm working on now, and wait to see if there's feedback. By the way, I do want to see my bot botflagged; after all, this is why bot flagging was invented (See: Rambot) Eugène van der Pijll 18:01, 17 August 2006 (UTC)[reply]
As I have to go slowly, by the way, I'm also doing South Tyrol at the moment. This province requires a lot of manual editting, so I'm doing that bot work very slowly. Eugène van der Pijll 15:39, 18 August 2006 (UTC)[reply]

This is a bot account, belonging to CyRoXX and is based on the Pywikipedia framework.

In the English version of Wikipedia it will be used to correct interwiki links. For this task, I run it in "autonomous mode". No critical edits should be expected from this mode, so I only supervise its interwiki link actions from time to time. When I check the intwerwiki links in the german articles once and find something to correct in en:, the possibility to correct it myself via bot saves time and mainly ressources, because you don't have to wait for another bot in the English version of Wikipedia.

At the moment, CyroBot is a registered bot on de: where he also performs other tasks, primarily text replacements. As I'm not that much online here, I ask you to inform me here when bot status is given (or not). Thanks. --CyRoXX 09:41, 30 July 2006 (UTC)[reply]

What kind of link corrections does it make. Does it ever add links?Voice-of-All 01:11, 31 July 2006 (UTC)[reply]

I need diffs on this one -- Tawker 01:14, 31 July 2006 (UTC)[reply]

Is this bot still being proposed or not?Voice-of-All 21:32, 7 August 2006 (UTC)[reply]

Awaiting Operator Response

  • What: The purpose of this bot is to do WikiProject maintenance for whatever WikiProject thinks its services are useful. WikiProject Anime and Manga will be the first beneficiary if this bot is approved. Of course, all its actions will have to be approved by any WikiProject before it is used on that WikiProject, as while in many cases they may be very useful, some WikiProjects may have different sorting systems that this bot should not mess with.
  • Why: I have noticed that WikiProjects run into a few problems in terms of raw sorting. First of all, there are often a huge number of stubs (hundreds or thousands) that belong in the category of a WikiProject (with the appropriate template in the discussion page) but have not been included. A bot could easily do this. There are other similar things, of course, that the bot could also do.
  • Exactly What:
    • Put all "anime and manga stubs" not in WikiProject Anime and Manga into Wikiproject Anime and Manga as Stub-Class.
    • If WikiProject Anime and Manga lists a stub as anything but stub-class, remove the stub tag as long as the article is more than 1000 characters long. If it is shorter, do nothing, as it could be an actual stub. The 1000-character cutoff isn't an arbitrary cutoff, but simply a safety feature to avoid a really short article that was accidentally listed as a non-stub in the article rating scale to stop being called a stub.
    • For all projects with their own rating system, turn all GA-class articles in those projects that aren't listed as GA or better into GA-class articles under that project (this will start with WikiProject Anime and Manga only).
  • It uses PyWikipedia Framework. It won't be a server hog because it will be run manually (i.e. only when I tell it to), so instead of patrolling recent changes for hours on end it will simply patrol the categories when told to, probably weekly or every few days.
  • There are thousands of articles here that need proper sorting, and a bot in my opinion is the best way to do it. In addition, if there is some sort of mistake (i.e. a B-Class article that's currently a stub that gets unlisted as a stub when it shouldn't be), it isn't devastating: a 99% success rate would add far more good than it would bad, and I highly doubt it would have any real failures, unless some fool ran around rating Stubs as B-Class articles. Dark Shikari 10:26, 18 July 2006 (UTC)[reply]
I don't see a problem, as long as it's being sponsored by a Wikiproject, and the exact details of what it is going to do are approved here in advance. We don't generally give a "whatever you need it to do" clearance; you tell us specifically what it will do, and we approve that. If you add something, you drop a note here saying what you're adding, and we say yea or nay. I see no problem why it can't have a trial period once it's set up and has a specific list of tasks. Essjay (Talk) 00:41, 19 July 2006 (UTC)[reply]
Its not really "sponsored" just yet: I'm asking what people think of it on the WikiProject talk page. So far I've received no disapproval--they've also made suggestions as to what I can and can't do in regards to sorting using the bot. You can find the discussion so far here. How much approval should I look to get before I set the bot into motion? Dark Shikari 01:09, 19 July 2006 (UTC)[reply]
Dear God no!!!! You don't seem to realise that stubs and Stub-Class articles are two completely different things! The terminology is admittedly extremely confusing (and the sooner something is done about it, the better). Also you clearly don't understand that length is only a minor consideration when it comes to working out what a stub is. An article swith one line of text followed by a list of 50 examples, or one line of text followed by a large table, is definitely a stub and likely well over 1000 characters. This is why stubs are sorted by hand, rather than by some automated method. Having the bot run in this way could well reverse much of the work of the Stub sorting wikiproject. Grutness...wha? 01:29, 19 July 2006 (UTC)[reply]
(Just for clarity) Hence why I said "as long as it's being sponsored by a Wikiproject, and the exact details of what it is going to do are approved here in advance." My take would be that the Wikiproject people would be quick enough to know what should and shouldn't be done, and that as long as specific changes are set out here before they are done, there is plenty of time for approval. Just saying that for the sake of clarity on my earlier comments. Essjay (Talk) 01:42, 19 July 2006 (UTC)[reply]

I have similar concerns to Grutness. The whole concept of a "stub class article" is poorly defined at best: at a deletion discussion on one such category, one WP1.0ist argued to keep them separate specifically on the grounds that "Stub class" is not the same as "stub"; another wanted to keep, on the basis that they were essentially the same. This needs to be much more clearly defined before a bot goes around making sweeping changes based on an assumption one way or the other. I see no evidence of support for this at the indicated Wikiproject, and the following comment sounds like opposition, or at least a reservation, to me: "It probably should go case-by-case (which I guess isn't what you want to hear for a bot, huh)." I'm especially opposed to stub-tag removal by bot; if a B-grade article is still tagged as a stub, there's clearly been a snafu someplace, since appropriate tagging and categorisation should be required to get it to that standard: much better to detect these by category intersection sorts of report generation, and have someone fix them manually. Stub-tagging might be more reasonable, but again, it's been specifically claimed by one person that some "Stub class" articles are not stubs, so this would require clarification and refinement; it would also have to be done very carefully to make sure that the "Stub class" tag, and the "stub" tag have the same scope by topic, and that no more specific tag applied instead. (For example, doing this to apply a single stub type to the whole of WPJ Albums, or WPJ Military History, would be a disaster.) Alai 04:06, 19 July 2006 (UTC)[reply]

I hope nobody minds my ill-informed opinion, but to incorporate Alai's suggestion into the bot sounds like a benefit. Without doing any editing, the bot could see how many articles have been snafu'd in such a manner and generate a report to somebody's userpage. It almost sounds like this should be a generic bot that Project leaders could download and configure for their specific purposes, or you could keep it proprietary and write all those little sub-bots yourself. ;) Xaxafrad 04:20, 19 July 2006 (UTC)[reply]
Well I had been informed by many that stub-class articles and stubs were in fact the same thing, and that anything above a stub-class article could not possibly be a stub. I had also been told that anything under the category "anime and manga stubs" most certainly belongs in Wikiproject Anime and Manga as a stub-class article. So who's right? I'm now getting confused... Dark Shikari 09:48, 19 July 2006 (UTC)[reply]
And in addition, wouldn't the most recent assessment of an article always be trusted? It seems as if most of the stub categorization was made months or even years ago, and hundreds of articles are still marked as stubs that have expanded far more since then, and have much higher ratings within the WikiProject. Are you saying that WikiProject ratings are totally useless and cannot be used to justify removing a stub tag? The WikiProject members would disagree. Dark Shikari 10:00, 19 July 2006 (UTC)[reply]
Oh, and also, is there anything wrong with simply having the bot shove all the Anime and Manga Stubs into the Wikiproject? There's a huge number that aren't part of the WikiProject, and nobody seems to have a problem with assigning them to the Project. Dark Shikari 13:03, 19 July 2006 (UTC)[reply]
If "Stub class article" and "stub" really are the same, then we're back with my original concerns, i.e., why have two sets of categories for the same thing, thereby duplicating work, and causing exactly the sort of inconsistency described? See for example this discussion, in which the matter is made... well, rather opaque, actually. One claim seems to be that a "long but useless" article would be "Stub class", but not a "stub" (though personally I would say that a textually long article can still be a stub, so even that's far from clear either way). It's possible one implies the other, but not vice versa, for example. No, I'm not saying assessment ratings are totally useless: are you saying stub tags are totally useless? Members of WP:WSS would disagree. An automated rule that simply assumes which of the two is incorrect is highly problematic, either way. An assessment that ignores the presence of a stub tag is either a) using different criteria for the two, or b) failing to leave the article in a state consistent with said assessment; a person has made them inconsistent, and a person (same or otherwise) should make a judgement as to how to make them consistent. People are stub-tagging things all the time, I see no reliable basis for assuming that the assessment rating is either more reliable, or more recent. I have no objection to the application of WPJ-tags by bot, if the associated WPJ has expressly agreed to the basis this is being done on, in their particular case. Alai 14:58, 19 July 2006 (UTC)[reply]

Thanks for the input, Alai. I've made a revised set of things the bot could do:

    • Put all "anime and manga stubs" not in WikiProject Anime and Manga into Wikiproject Anime and Manga as Stub-Class. There's no way for the bot to figure out if they deserve a higher rating than Stub-Class, so its fair to start them off there.
    • For all projects with their own rating system, turn all GA-class articles in those projects that aren't listed as GA or better into GA-class articles under that project (this will start with WikiProject Anime and Manga only). Do the same with FAC articles that aren't listed as FACs under the WikiProject system. Dark Shikari 15:15, 19 July 2006 (UTC)[reply]
      • That seems completely fine, no objections to those tasks. Obviously in each case consultation with the wikiproject, and confirmation with them of the scope of the articles they're "adopting" thereby would be indicated. Alai 02:59, 21 July 2006 (UTC)[reply]
      • Seems fine to me as long as there is agrement at the project.Voice-of-All 17:14, 29 July 2006 (UTC)[reply]





  • Before I run the bot, I want to see what Wikipedians think of another idea I had in mind (possibly a second thing the bot can do). How about a routine that checks through Wikipedia articles and fixes redirects: i.e. if an article links to "Blue", when the actual article is "Blue (colour)" and there is a redirect between the two (not a disambig), the original article will be changed to link to "Blue (colour)." The appearance of the page will remain unchanged: only the link in the code will be changed. This would probably slightly lower the load on Wikimedia servers by lowering the number of redirects. In addition, its always struck me as somewhat odd to have articles link to redirects, as redirects make more sense as methods of catching various user-inputted article names and sending them to the right article. This especially applies to older articles that link to a main article which has now been moved. What do you all think? Dark Shikari talk/contribs 19:41, 31 July 2006 (UTC)[reply]
  • Doesn't seem such a good idea to me. There's often been complaints about "needless" edits to eliminate redirects, and the opinion has been expressed that they cause more load issues than they solve, due to edits being much (much, much) more expensive than page fetches. Alai 17:36, 4 August 2006 (UTC)[reply]
  • Figured. Seems like a reasonable idea, but below the surface there are hidden problems, I guess. Or else someone else would have already done it ;) — Dark Shikari talk/contribs 19:05, 5 August 2006 (UTC)[reply]
  • Actually, it's a feature I've been working on quite some time ago. I think that a massive fixup of links on a page is a good idea. My MiszaBot has even done a few such edits in the past (examples on its userpage). As soon as I rewrite the bot to use the pywikipedia platform and make sure it conforms with Wikipedia:Redirects, I'll propose it here, since I now have the ability to set it up as a regular service. Misza13 T C 21:00, 5 August 2006 (UTC)[reply]
Now that this is under the Awaiting Operator Response category, I'm going to talk with the Wikiproject folks to ensure that they think this bot is fine. — Dark Shikari talk/contribs 13:36, 16 August 2006 (UTC)[reply]

This will be a semi-automated account, but in full control by me. It will take yesterday's category from proposed deletion, check the articles in that category, and apend a message similar to this:

An article that you created, [[:{{{1}}}]], was proposed for deletion, probably yesterday. Please review the policy on proposed deletion and feel free to comment on the article's talk page. If no contest is made, the article will be deleted in four days from today.

I just want to make the edits under this account because they will be repetitive and I want to do them quickly. The bot should run daily starting at 00:10 UTC. User:AmiDaniel is scripting a process for me and it will probably be integrated into AWB. He told me to get a bot account for this.--Kungfu Adam (talk) 23:28, 4 August 2006 (UTC)[reply]

I really like the idea. If AmiDaniel is working on it, obviously the technical end is pretty much covered. My only concern is that this is going to greatly increase the number of unexplained prod tag removals (and thus significantly increase backlog at WP:AFD. Any thoughts on that? alphaChimp laudare 19:28, 5 August 2006 (UTC)[reply]
I like the idea because I am tired of getting messages, WAAA you deleted my article, why didn't you tell me? I spammed about 50 talk pages yesterday with this message and I quickly got burned out, so I want some sort of automated process.--Kungfu Adam (talk) 20:45, 5 August 2006 (UTC)[reply]
Is there actually a need for this?
  • Is there anything to suggest prod is deleting lots of good articles which shouldn't be deleted
  • Why choose to tell the creator, WP:OWN they don't own the article, what if the creator was of a stub but the bulk has been contributed by others?
  • Isn't part of the purpose of prod to remove non-speedyable stuff, which is nevertheless questionable. If no one is actually watching/interested in the article why create an artificial interest from the intial creator.
  • Articles can always be undeleted if prod has failed for some reason.
  • Is this actually a real problem, i.e. are there people complaining about lack of notification?
Just some thoughts --pgk(talk) 21:05, 5 August 2006 (UTC)[reply]
I think pgk has it right here: If they aren't watching, why tell them? I've done prods a few times, and I've had the resulting "Why?" notes; I point them at the prod policy and say "Apparently, you weren't watching." I'm not sure bot notifications will do much to help; I'm positive they will cause prod-related edit wars. Essjay (Talk) 07:29, 6 August 2006 (UTC)[reply]

Alright, well, I went ahead and finished up the script tonight. Only took me about an hour and a half to do, so should the bot not gain approval, 't won't be all that much of a loss to me. I did a test run on test.wikipedia, whereby I used Category:Proposed deletion as of 2 August 2006 and posted the messages on user talk pages on test. You can see that test run at test:Special:Contributions/AmiBot. The basics of how it works entail loading up the CAT:PROD of four days one day previous using query.php and then again using query.php to fetch the last 20 edits by unique authors to each of the articles in that category. If it returns exactly 20 contributors, then, under the assumption that the original contributor had little to do with the article, that editor will not be added to the notify list. Also, as I found many of the articles in the test cat were old enough to have been created by anons, if the original creator is an anon, that user will be excluded as well. Generating the list of original contributors (thanks to Yurik's fabulous Wiki API) takes only about 30 seconds, and then editing at about 5 edits / sec (the throttle can also be adjusted faster and slower) leaves around an average of 15 minutes to notify all the contributors. I don't have it set for scheduled runs currently, so Adam will have to know to run it daily, though it won't require too much work on his behalf. Anyway, I think pgk's concerns are quite valid, some of which we may be able to address (i.e, per concerns of WP:OWN, it could notify all contributors to the article, which would naturally result in longer runs and random stub-sorters being informed, but may be a possible solution). I do, however, believe that propper notification of proposed deletions would be advantageous to everyone involved, including the poor admins who get lambasted by the malinformed and confused original creators for deleting uncontested prods (as I have been many times), and it would seem to me that PROD's purpose is not to try and slip deletions by unsuspecting editors but to see if there is any real need to discuss an article before deleting it. In any case, I would like to give it an actual trial run for a few days on enwiki to see what happens, and if it results in mass deprodding, etc., we can reconsider. Source code readily available upon request (just please let me know in advance so I can clean it up a bit =D). AmiDaniel (talk) 07:44, 6 August 2006 (UTC)[reply]

I also just ran another trial on test.wikipedia using Category:Proposed deletion as of 3 August 2006, which had slighlty more articles and were slightly more recent. It went fine as well. Let me know of anything that may need to be changed. AmiDaniel (talk) 09:09, 6 August 2006 (UTC)[reply]
I value all of these concerns, and believe they are valid, but if anyone wants to know if people complain or not, they do. All you have to do is look at my talk page, and look in the archives. However, the number of complaints are very small in comparison to the deletions that I do. But there are some good articles that pass by. Yesterday, when doing Prod deletion, I removed a tag because I believed it was encyclopedic. These are my thoughts.--Kungfu Adam (talk) 11:30, 6 August 2006 (UTC)[reply]
A further comment to this would be are we trying to patch up a problem in the prod process with this. If this is a real issue, shouldn't prod itself mandate the adding of notices to creator/whoevers talk page? --pgk(talk) 20:15, 6 August 2006 (UTC)[reply]
I believe that there is a problem with the current prod process, be there mention of it in WP:PROD or not. The entire purpose of proposed deletion is to see if there are objections to deleting an article--that is, if there needs to be community discussion before deleting an article. Some believe that simply adding the tag to the article is enough to see whether or not anyone objects; however, I would disagree. A few scenarios: One, I have over 3,000 articles on my watchlist. The odds that I would happen to notice someone randomly adding a prod tag to an article that I've worked on are very slim, though I may full well object to one of those 3,000 articles being deleted. If someone, bot or user, leaves a kind message on my talk page about an article I've created or worked on, it then will provide me with my formerly unavailable object to contest the article's deletion before it is actually deleted, much as notices from OrphanBot regarding nsd's and bad licenses allow uploaders the opportunity to address concerns that may lead to their file's deletion. Two: Many new users do not understand how to use watchlists, check deletion logs, etc., and they often don't check up on the articles they create and/or work on extensively. When the article then one day turns up missing with no notice or apparent reason, they feel confused and frustrated and either leave the project silently or lash out in rage at anyone they can blame for the article's deletion; thus, by notifying the contributors to these articles that are proposed for deletion, we lessen the likelihood of biting the newcomers. As such, I do believe that there are problems with how prod works currently, and I think that Adam's idea may well serve as a cure for it. I'm open to suggestions about tweaking the bot to perhaps post messages to anyone who's contributed more than one edit to the article, etc., and perhaps expanding the message to point to relevant policies and guidelines, such as WP:DEL, WP:N, etc. I think it's also worth considering that this bot (in its current form) will only run for about twenty minutes a day and make fewer than 200 edits a day with little to no apparent load on the servers and a much slimmer chance of it going haywire than many of the other bots we have running. So since the sacrifice and risk we're taking by running this bot is minimal, it would seem to me that if there is any benefit to running (which I believe there most certainly is), then it should be allowed to run. I'd also like to point out that this bot will pretty much just be doing what Adam is already doing and will likely continue doing regardless of the outcome, only faster and more efficiently. Anyway, another soliloquy from me hath come to an end =D. Please let me know where you believe my conclusions are mistaken. AmiDaniel (talk) 04:44, 8 August 2006 (UTC)[reply]
OK, you seem to suggest there *might* be a problem, which is why I originally asked "Is there anything to suggest prod is deleting lots of good articles which shouldn't be deleted". And we still don't seem to saying there is in reality a problem. I would suggest that if there is a problem with prod it should be discussed on the PROD talk pages, or VP first. If the conclusion is that notification is a good thing then a bot is quite possibly a good solution, what I don't think we should be doing is declaring a solution when it isn't clear there is actually a problem, or if discussed a better solution might be found. There is a large difference with the tagging and warning of image, only one person can have uploaded the current version of an Image and more importantly until very recently it was terminal, with the recent changes to allow undeletion it may actually warrant a review of how we handle image deletion and may deemphasise the warning aspect. I would also somewhat disagree with your description of PROD, the idea of PROD is still fundamentally to get rid of junk, only as a more lightweight process, the admin dealing with a prod article at the end of the period should NOT be blindly deleting, that would be process idiocy. The admin should contest deletion on anything he doesn't feel should be deleted in that manner, of course noting that removal of PROD does not imply he has to AFD it. --pgk(talk) 18:28, 8 August 2006 (UTC)[reply]
Did the note at the bottom really mean to put new messages here? If so, I think the bot is a good idea. I did not create, but had contributed to a list that was PRODed in four days and without any notice. (Deleted 19:02, 22 July 2006 by Kungfu Adam (talk) Prod placed by 16:51, 18 July 2006 Recury) It was on my watchlist, but went under my radar. This was the first I had dealt with a prod. Doctalk 12:13, 14 August 2006 (UTC)[reply]
The bot is designed to contact the creator, not all contributors. Contacting all contributors would be a strain on server resources, and overkill.--Kungfu Adam (talk) 16:28, 14 August 2006 (UTC)[reply]
Yes, I realize that I would still have "lost out" on this one, but from the experience, I can identify with how frustrating it would be to have created an article and not even have received notice that a prod had been placed. Doctalk 16:39, 14 August 2006 (UTC)[reply]
I like the idea, but you might consider a couple of variants. One would be to notify those who have the article on their watchlist. Another would be to notify (say) the most recent five (or whatever) contributors.
I don't agree that the measurement of the problem is only those who complain. First, there are an unknown number of editors who don't notice, but would complain (and when they do notice, it's not timely), and an unknown number who do notice but decide it's too much trouble/too little likelihood of change to bother complaining. Second, the bot improves the perception of fairness in wikipedia; that's important to building a sense of community.
And personally, I'd like to see this for AfDs too - again, either for those who have the article on a watchlist or for recent contributors. John Broughton 23:16, 14 August 2006 (UTC)[reply]
John, perhaps that is a good idea, but it is impossible to notify editors that have something on their watchlist, because the watchlists are not public information like almost everything else on Wikipedia is.--Kungfu Adam (talk) 10:49, 15 August 2006 (UTC)[reply]
This still seems to be to be the wrong place to be discussing this. If there are problems with process from AFD, PROD etc. It should be discussed there first, if the solution proposed there fits a bot then by all means we can propose/use a bot, but this isn't the right place to be determining new process/policies. I'll DENY this for now pending discussion within the PROD etc. processes --pgk(talk) 11:04, 15 August 2006 (UTC)[reply]
Well, I won't strongly debate your denying the BRfA, simply as I do recognize there are a few things that will need to be worked out with the bot; however, I do completely disagree with your reason for denying it. It would seem this discussion has been had, and the result was in favor of notifying contributors. To quote from WP:AfD: "It is generally considered civil to notify the good-faith creator and any main contributors of the article that you are nominating the article. To find the main contributors, look in the page history or talk page of the article and/or use TDS' Article Contribution Counter. For your convenience, you may use {{subst:AFDWarningNew|Article title}} (for creators who are totally new users), {{subst:AFDWarning|Article title}} (for creators), or {{subst:Adw|Article title}} (for contributors or established users)." To quote from WP:PROD: "Consider adding the article to your watchlist and letting the article's creator know that you have tagged it. You can use {{subst:PRODWarning|Article title}} ~~~~ for this." So anyway, I don't know what discussion you want me to have--perhaps I should discuss not changing our current policies and guidelines? Until now I've never even thought that it could conceivably be a bad idea to notify contributors that their articles may be deleted, and it would seem that consensus has quite consistently been for the idea that leaving such notices was civil, decent, and beneficial to the deletion process. As there has long been agreement that letting everyone involved know about proposed deletions would be ideal, it would seem the only question that remains is whether or not posting these notices should be automated or not. A BRfA seems to be the perfect forum to discuss whether or not to automate a generally accepted process, but apparently not. So tell me, where should we go now? In any case, Adam and others who recognize the benefits of doing so will continue, as they have been doing, notifying contributors in a highly inefficient and tedious manner, but perhaps one day we'll be able to let the bots do the monkey work and let the humans do the encyclopedia-writing. AmiDaniel (talk) 11:56, 15 August 2006 (UTC)[reply]
Well I guess I'm just showing my ignorance of what is currently written in our processes.
"It would seem this discussion has been had, and the result was in favor of notifying contributors", we've had a discussion here on a request for a bot, my point is that this is a narrow audience and not the place for a general discussion on how we should implement process and policies relating to deletion. Someone interested in deletion processes probably doesn't come here looking for such debates. If that discussion has been had elsewhere please link to it. This is similar to the situation regarding tagging for deletion of empty categories, an apparently innocuous process which seemed to upset a few people.
You say it is our current process and a generally accepted process, then why isn't it being done by people when they tag articles? Our process/policy is what we actually do, not what is written down since that written expression by nature always lags reality. To say it is broadly accepted when very few actually do it (hence the need for a bot) shows clearly that either there isn't a broad acceptance of this standard, or that people find it a chore so skip it as optional, in the latter case a bot may indeed be a good idea, but unless we find out why people don't do it we simply can't tell which it is (or indeed if there are other reasons people don't do it)
Regarding why it might be a bad idea is a pretty open ended question, as suggested by other above one reason is that it might end up making PROD ineffectual in getting rid of junk and just push that back onto AFD. Another is I have suggested that WP:OWN comes into effect, the original author has no special relationship or say over the content. And I'm sure others may come up with other reasons (Off the top of my head those who regularly create pages for anons via a request for page creation maynot care to be regularly spammed)
The discussion I'm suggesting you have is in the relevant places for PROD (and if extended to AFD there too), such that a broader input can be gained into if the tagging is a good/bad idea, if a bot doing it would help etc. etc. I cannot see why there is any resistance to this, since you seem pretty convinced it is a no brainer. --pgk(talk) 12:31, 15 August 2006 (UTC)[reply]
As a slight addendum, I should note that the list of thoughts I gave above was the result of a brief discussion with several people, this is why I feel there may actually be more disagreement on this than you appear to think (I could still well be wrong) and so a more general discussion on the basic issue might be of longer term benefit. --pgk(talk) 12:47, 15 August 2006 (UTC)[reply]

Requests to add a task to an already-approved bot

Alphachimpbot, 5th Function: Architecture Category

Per this bot request, I'd like to add the {{architecture}} tag to pages within the Category:Architecture. alphaChimp laudare 23:23, 15 August 2006 (UTC)[reply]

Awaiting approvals group response

FFBot, new fuction

As per [2]. — FireFox (talk) 18:45, 15 August 2006

FireFox has gone on a leave of absence requset permission to tske over this Task with my bot Betacommand 15:56, 18 August 2006 (UTC)[reply]

Awaiting approvals group response

FFBot, third fuction

This will be the third function if the below request is approved. I'd like approval for FFBot to help out with WP:CfD – I've noticed there is often a slight backlog here and I see no problems with having another bot to help out. If approved, it will move all involved pages into a different category per the discussion result, and then I'll delete the old category. — FireFox (talk) 18:08, 13 August '06

Give a cpl diffs pls -- Tawker 19:01, 14 August 2006 (UTC)[reply]
[3] [4] [5]. Is this enough? — FireFox (talk) 19:10, 14 August 2006
I'll just post a few more just in case: [6], [7], [8]. — FireFox (talk) 20:14, 14 August 2006
Yep, go ahead, that looks fine -- Tawker 02:13, 15 August 2006 (UTC)[reply]
Thanks. — FireFox (talk) 10:05, 15 August 2006

FFBot, second fuction

Quite simple: replacing {{cleanup}} with {{cleanup-date|...}}. There is at least one other bot which does this, but in my opinion, the more bots there are to do the job, the more efficiently the job gets done. — FireFox (talk) 20:38, 9 August '06

I see no problem with it. Want the AWB settings file? If so, you'd be running the same code as another bot approved for the function. alphaChimp laudare 02:11, 12 August 2006 (UTC)[reply]
I don't mind, it's not hard to set up (I think). — FireFox (talk) 11:04, 12 August '06
Straight forwad enough good ahead. --pgk(talk) 19:46, 14 August 2006 (UTC)[reply]
Thanks. — FireFox (talk) 20:14, 14 August 2006

User:DumbBOT, fifth function

Some articles and images are tagged {{copyvio}} but are then not listed at WP:CP. I'd like approval for User:DumbBOT to perform the second step for these incomplete nominations. This is done by loading Category:Possible copyright violations, waiting 120 seconds (to give time to complete the noms), loading WP:CP, and comparing the list of articles/images in the category and the list of articles/images linked from the copyright main page. The resulting lists are then posted as in [9] and [10]. (Nothing is posted if the list to post is empty.) I'd run this bot manually. (Liberatore, 2006). 12:50, 1 August 2006 (UTC)[reply]

How often would this be run and could the wait be incresed to 300 seconds?Geni 14:39, 1 August 2006 (UTC)[reply]
As far as I can see, there is probably no need to run it more than once in two/three days (I am not planning to run it scheduled.) I could increase the delay to 300, or also to 600 secs; there is no problem with this. (Liberatore, 2006). 14:50, 1 August 2006 (UTC)[reply]
600 would be overkill I think 5 minutes is enough.Geni 15:15, 1 August 2006 (UTC)[reply]
Wikipedia:Copyright problems already has a large backlog. Will adding more to it help resolve things? — xaosflux Talk 01:49, 9 August 2006 (UTC)[reply]
Not adding copyvio-tagged pages where they belong won't fix things, though. Titoxd(?!?) 01:56, 9 August 2006 (UTC)[reply]
(To Xaosflux) Pages tagged as copyvio are still pages to be processed and possibly deleted, whether they are listed in WP:CP or not. Otherwise one could delete all links from WP:CP and claim to have eliminated the backlog.
I agree that a bot that helps on this backlog might be useful. The only possibility I see is a tool for comparing the wikipedia page with the alleged source. However, this has still to be checked manually, plus there are pages that are a merge of two or more sources.
Incidentally, I'll be on vacation from Sat for a couple of weeks. Is it ok if I run this bot once today or tomorrow and finish the trial (if authorized) when I return? (Liberatore, 2006). 11:41, 9 August 2006 (UTC)[reply]
I forgot to mention that I posted a link to this discussion at Wikipedia talk:Copyright problems#DumbBOT days ago, and nobody objected, beside the comment above on the delay (Liberatore, 2006). 11:47, 9 August 2006 (UTC)[reply]

Grafikbot, MILHIST article tagging

I would like Grafikbot (talk · contribs) to tag military history related articles with the {{WPMILHIST}} template for assessement purposes as defined in the WP:1.0 program.

A complete automatic tagging bot is still out of reach, so for the time being, a limited tagging will be executed as follows:

  • Every article with {{Mil-hist-stub}} stub tag and related tag (list available at WSS) as well as with {{mil-stub}} and below (list available at WSS) is considered as a military history article and thus subject to tagging.
  • The list from Wikipedia:WikiProject Military history/New articles will also be processed.
  • The talk page of each article is tagged with the {{WPMILHIST}} prepended to the talk page (even if the talk is empty).
  • The run is repeated say one or two times a month to make sure that new stubs get properly tagged.

Note: a rather lengthy debate took place on WP:AN a few weeks ago, and a consensus emerged that such a tagging was desirable for the whole WP project. Obviously, a bot can't tag everything, but I think it just can handle this one. :)

Can someone approve this please? :)

Thanks, Grafikm (AutoGRAF) 15:09, 21 July 2006 (UTC)[reply]

  • This somewhat touches on some of the same issues as discussed in relation to Dark Shikari Bot, but I see no problem as such. The scope does seem rather wide, though: {{mil-hist-stub}} obviously makes intuitive sense, but does the wikiproject really want to "adopt" the whole of {{mil-stub}} (i.e. corresponding to the whole of Category:Military)? Perhaps when you're approved for a trial run, you might start off with just the former, and consult somewhat on the latter. Alai 04:48, 23 July 2006 (UTC)[reply]
    • Well, aside from fictional stuff (which people really shouldn't be using {{mil-stub}} for anyways, I would think), we've already adopted basically all of Category:Military already. Kirill Lokshin 04:54, 23 July 2006 (UTC)[reply]
      • Then it's not very well-named, is it? (BTW, you rather said the opposite when a "Stub-Class articles" category was being discussed for deletion, that there was no single hierarchy to your WPJ's scope...) At any rate, if there's any scope whatsoever for "false positives", it's not the best place to start. Alai 05:10, 23 July 2006 (UTC)[reply]
        • Well, no; the project's scope is actually broader than merely what's in Category:Military ;-) As far as false positives, I don't know how best to handle that. (What's the danger of having a few extra articles tagged, though? These are only talk page tags, and tend to be removed from articles where they don't belong with minimal fuss.) Kirill Lokshin 05:14, 23 July 2006 (UTC)[reply]

For your collective information, I note that there are 11,216 articles in or under {{mil-stub}}. That even misses a few, since for some incomprehensible reason, aircraft are largely split by decade, rather than into military and non-. There's 3,374 that are specifically "military history". Let's be as sure as possible these are all within the Wikiproject's scope before getting carried away with this (especially the "non-historical" ones). Alai 07:40, 25 July 2006 (UTC)[reply]

  • Well, I am leaning against a bot. I have just been cleaning up some 75 articles in cartridge category that had the WPMILHIST banner on the talk page that I assume were done by a bot. If the article was read it clearly stated that it was used for sporting or hunting with no military referenced at all. Article should be read and assessed at the same time.--Oldwildbill 10:45, 25 July 2006 (UTC)[reply]
    • There are no tagging bot here that I'm aware of, at least not today. -- Grafikm (AutoGRAF) 17:00, 25 July 2006 (UTC)[reply]

I just took a quick run through the various stubs involved here. The only one in which I found articles not within the project's scope is {{weapon-stub}} and its children, as there are some hunting-related articles there. Hence, we should not automate tagging for that tree. As far as I can tell, however, all of the other children of {{mil-stub}} are reasonably safe to tag by bot. Kirill Lokshin 04:00, 26 July 2006 (UTC)[reply]

So, are there any other opinions on this point? The project discussion on this issue wasn't very active, but there weren't any objections to this tagging proposal. Kirill Lokshin 18:27, 9 August 2006 (UTC)[reply]
Looks like there is a heavy backlog on this page currently... -- Grafikm (AutoGRAF) 21:47, 9 August 2006 (UTC)[reply]

Bots in a trial period

This is an AWB bot that will be used for converting {{Wikify}} and {{wikify}} to {{Wikify-date|{{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}. Currently I use the account for spellchecking, to seperate my AWB edits, however every edit on the account has been checked by me - I opened the account because of a suggestion on the AWB page.

Currently all it will do is to convert the above tags, using AWB's find and replace, however it will be unassisted except for being started once per day. If the bot is approved I will request AWB approval for a bot and set it to 'Auto' mode.

The bot currently has AWB approval as a user only. I have converted a few tags using AWB by reviewing each edit's 'show changes' myself and checking articles afterwards (see [11] for contrib list). To demonstrate it I will apply for AWB approval as a bot when required. I will not suppress the 'using AWB' messages.

The bot will run for a short period each day (depending on the number of articles marked with '{{wikify}}') for as long as it is needed.

The bot will make AWB edits on the english wikipedia only. I (Draicone (talk)) will be the sole operator of the bot.

The bot will give those of us at Wikipedia:WikiProject Wikify and Category:Articles that need to be wikified a better idea of the number of articles to be wikified in the month. I understand that Pearle is currently serving about the same purpose. DraiconeBot will simply help clear backlogs as they come so that there is not a major queue waiting for Pearle each day. Server load will be minimal, I will aim to keep server load minimal with 4 edits per minute (approx.) using AWB delay, I can adjust this if requested. --Draicone (talk) 23:13, 11 August 2006 (UTC)[reply]

One week trial run approved, keep edits down to 2-3/min during tial, limit run to 100 articles. — xaosflux Talk 22:39, 13 August 2006 (UTC)[reply]
Alright, had AmiDaniel give AWB bot approval and ran a trial run of just over 50. Worked perfectly besides a couple of things I forgot. Summary:
  • More or less all wikify tags were added today anyway, so checking when the tag was added isn't an issue and AWB does the trick.
  • Sample diffs: [12], [13], [14], [15], [16], [17], [18], [19]
  • Points to remember: I didn't set 'Mark all as minor' at the beginning, only remembered it later on. Should these bot actions be minor edits? Also, the bot made one bad edit, diff link [20]. I forgot to filter out non-mainspace articles from the list, I will remember this for next time. I reverted this edit almost immediately.
Nothing of great significance, so I've decided to end the trial run at these 60 or so articles. Please tell me if you would like to extend the run. --Draicone (talk) 09:13, 15 August 2006 (UTC)[reply]

Awaiting approvals group response


A long time ago, I ran User:The Anomebot, implemented in Python, which was one of the first Wikipedia bots. I now want to register a new bot account for running a simple bot, based on the Wikipediafs Debian package and a simple Python editing script, that will add geotags to allready-existing geographical articles without geotags. As part of this work, I have compiled a list of just under 10,000 new geotags, which has already undergone spot checks for validity. The bot will check for existing tags during its run, and will not attempt to replace any existing tags.

The following data is the result of taking Wikipedia's category links and the public domain NIMA GNS data, and rubbing vigorously. With quite cautious checks applied to both datasets, this gives an unambigious location for 12660 out of a possible 28628 (44%) articles about non-US cities, towns and villages.

The results are sorted by country, then place, and binned into four files. They have also been compared to the data in Koordinaten_en_CSV.txt, and labelled by whether they are new coordinates (NEW), or duplicate coordinates already in articles (dup), and, if so, whether they are exact duplicates, or if not, roughly how many km out they are. (The distance calculation uses several approximations, so treat it only as an order-of-magnitude figure).

Where this data differs from the existing Wikipedia data, the new data has been found to be correct in almost every case: where the data was in error because of bugs in the list compilation, I fixed the bugs, and regenerated the output data to remove any similar errors.

The bot is intended to run at a limited edit rate, and is intended to self-check its edits independently of Wikipediafs, and to stop if its edits are not saved, or differ in any way from the intended content: it should thus automatically stop if blocked, or if the Wikipediafs or the Wikipedia servers are malfunctioning. -- The Anome 02:16, 11 August 2006 (UTC)[reply]

One week trial period approved, please limit speed to no more than 2 edit/min; and limit the test run to 100 pages or less. — xaosflux Talk 16:05, 12 August 2006 (UTC)[reply]
Thank you. I've now completed testing this bot: I've checked a number of possible error scenarios, including restarts and auto-shutoff, and it seems to be working OK. -- please see [21] for the results. -- The Anome 11:35, 13 August 2006 (UTC)[reply]
Following Eugène van der Pijll's feedback, I've now made several more improvements, with a number of other filters being used to suppress edits in cases where geotags are already present either directly or through transclusion, and better placement of the tag within the article. I think I'm ready to run. I'd like to perform another limited test run of 100, just to make sure everything works OK, if you can approve that. If manual review of those 100 edits shows no problems, I think I'll be ready to run on the full dataset, subject to approval. -- The Anome 15:29, 13 August 2006 (UTC)[reply]
Additional 100 article run is OK, please post results here. — xaosflux Talk 15:59, 13 August 2006 (UTC)[reply]

I've now done another few small test runs, the most recent of which started at 23:44 with Aparan, and ended at 00:36 with Berry, New South Wales (with a slight tweak to the edit comment string in the middle for a couple of entries). The bot seems to be working OK in this run:

  • Articles with categories, interwikis and Unicode characters are handled OK, and the tag is added in the correct place: see Goris (as a side effect, the category and interwiki links are also sorted into the standard order after all other page content, and are sorted in alphabetical order)
  • Articles marked with a variety of pre-existing geodata template styles seem to be caught now: see Augusta, Western Australia, and Yerevan for examples.
  • Restart is working OK
  • Blocking the bot shuts it down as soon as it detects the write error
  • New feature: as requested by Eugene, the geodata tags now have region and feature codes, for example {{coor title dm|34|47|S|150|42|E|region:AU_type:city}}

Please let me know what you think, and whether you would like me to do any more testing. -- The Anome 23:46, 13 August 2006 (UTC)[reply]

Update: I've regenerated the input dataset using the latest dump data, and performed a few more bot edits to validate the new list, staying within my existing test allowance. I now have geodata available for more than 15,000 towns, cities and villages alone. I can easily do the same for other classes of geographic features later, using the same bot, and the same data-generation code, just by changing the GNS and category filtering parameters. Please let me know if/when I can proceed with adding the city data. -- The Anome 13:14, 18 August 2006 (UTC)[reply]

Looks very solid to me -- permissions is granted - do you want a botflag? -- Tawker 15:30, 18 August 2006 (UTC)[reply]

Awaiting approvals group response

This is a simple bot to clean up the huge userbox backlog Here. It will use AutoWikiBrowser and will only run when I give it the instructions.

For example it will replace {{User diy}} with {{User:Holek/Userboxes/Do It Yourself}}

It pretty much does the same job that Botdotcom bot does.

It is operated only by me, Andeh.

I could do this task on my current account but feel it would be better opening a separate account to do it.

When the backlog is cleaned up, I will request an administrator to block the account indefinitely. If I decide to give it another function, I'll request it here first before requesting the account to be unblocked.--Andeh 11:03, 9 August 2006 (UTC)[reply]

Ok, give us a few diffs and we'll approve it from there -- Tawker 05:24, 10 August 2006 (UTC)[reply]
Could you add the account to the AWB allowed list? Then I'll do a few to show you I know what I'm doing.--Andeh 15:30, 10 August 2006 (UTC)[reply]
Done -- Tawker 15:57, 11 August 2006 (UTC)[reply]
One week trial period approved, please limit edits to no more than 2/3 per min and limit the trial run to 250 replacements. — xaosflux Talk 02:11, 12 August 2006 (UTC)[reply]
Started, see AP.BOT (talk · contribs · logs).--Andeh 14:06, 14 August 2006 (UTC)[reply]
249 edits to user space;

talkcontribspage movesblock userblock logcount--Andeh 15:06, 17 August 2006 (UTC) [reply]

Hello;
I'm asking approval to run my bot, running Pywikipedia software. For now it will be only used in manually assisted mode for interwiki links on specific articles and, more frequently, to add interwikis using warnfiles from the Interwiki link checker, as shown on its user page. It is currently running on it.wikipedia with about 8000 edits. Thanks. --.anaconda 06:35, 8 August 2006 (UTC)[reply]

Does the interwiki checker depend on the Toolserver's replication lag? It is approximately two months out of date now, so it may not be ideal for your purposes. Titoxd(?!?) 07:47, 8 August 2006 (UTC)[reply]
No, it uses a list of articles with the same name on 2 languages, updated a few time a month. Someone checks if they deal with the same argument and then a bot adds the links. If the link exists will not be added. --.anaconda 13:56, 8 August 2006 (UTC)[reply]
Per it's edit history on .it this bot looks like it knows what it's doing. One week trial approved. Please restrict trial run to no more than 150 pages, and throttle edits to no more than 2/min. — xaosflux Talk 01:38, 9 August 2006 (UTC)[reply]
Thanks. I made 84 edits, most using warnfiles, so I think I'll stop here. --.anaconda 06:21, 10 August 2006 (UTC)[reply]
May I suggest maybe to use the interwiki.py instead of the warnfile.py for implementing Flacus' warnfiles. It is slower, but more secure. A lot of Flacus' warnfiles are not up-to-date anymore, even in the recent lists (week 31 for example) there are pages which have changed into redirects after the creation of the file. Interwiki.py will take this into account. In the case that there is a problem this program warns you and asks what to do instead of just doing its edit without further consideration.
For example this edit [22] could at the same time have copied all those links from en: to it: making it unnecessary for another bot to pass by and edit the page again. Interwiki.py would have done this itself. Thijs! 10:29, 10 August 2006 (UTC)[reply]
Thanks for your suggestion, never noticed the -warnfile on interwiki.py till yesterday :-). I used interwiki.py -warnfile yesterday for about 150 edits on it.wikipedia and seems good. --.anaconda 02:07, 11 August 2006 (UTC)[reply]
Trial count extended to another 150 edits, can you try again with your new settings? — xaosflux Talk 06:28, 12 August 2006 (UTC)[reply]
Sure, restarted. Cheers --.anaconda 15:30, 12 August 2006 (UTC)[reply]
I made another 144 edits, for a total of 234. So I sit here waiting for appreciated comments. --.anaconda 20:46, 13 August 2006 (UTC)[reply]
Another warnfile made additional 47 edits for a total of 281, so I can quite consider the trial ended. Here all its contributions. Comments, problems? Thanks. --.anaconda 11:31, 15 August 2006 (UTC)[reply]

Awaiting approvals group response


I am using this account with AWB in a semi-automated fashion to tag talk pages with WikiProject templates. This is a case where automation is of very clear benefit - we on the Wikipedia 1.0 team need these talk pages tagged for our article assessment drive, and the work is ridiculously simple for a bot but slow and boring for a human. I always test my regular expressions and settings on some test pages with a totally manual run first and the work is simple. Duration: As long as it takes. --kingboyk 12:13, 6 August 2006 (UTC)[reply]

Ok, go ahead and trial it, post some diffs and we'll take a look -- Tawker 00:58, 9 August 2006 (UTC)[reply]
What kind of diffs are you after? Or would you care to choose a few edits at random from the bot's history? That might be better: I have nothing to hide and it's all simple stuff :) --kingboyk 15:00, 11 August 2006 (UTC)[reply]
It's been a week, and I'm sure people are sick of seeing my bot's edits in recent changes. I've tagged masses of talk pages, received 3 barnstars for the work, had a number of enquiries which have all been dealt with, and no complaints. Let's get this bot bit then please! :) --kingboyk 13:25, 18 August 2006 (UTC)[reply]

I am working on developing a system for monitoring key administrative and editting related categories for the purpose of automatically identifying backlogs and other areas in need of attention.

An output prototype can be seen at User:Dragons flight/Category tracker.

The largest categories, those with 600+ entries, are scraped from Special:Mostlinkedcategories. The small ones are gotten by directly loading the category pages.

When completed, there will be a control page at User:Dragons flight/Category tracker/Config for adjusting which pages are tracked, how often they are monitored, etc.

Dragons flight 06:07, 31 July 2006 (UTC)[reply]

How often will this run? — xaosflux Talk 00:44, 2 August 2006 (UTC)[reply]
Well, one could make an argument that there are a few categories (e.g. Wikipedians looking for help, Requests for unblock, Wikipedia protected edit requests) where it might make sense to monitor them for changes several times per hour, but for the most part I was thinking from several time per day to once every few days. Special:Mostlinkedcategories only updates twice a week, so there is no point hitting that more often. I'm not actually sure how often would be most useful, which is why I was planning to build in some flexible control. Any thoughts? Dragons flight 02:14, 2 August 2006 (UTC)[reply]

Okay, the code has been more or less fleshed out. Please see: User:Dragons flight/Category tracker and associated pages for descriptions. Dragons flight 07:22, 8 August 2006 (UTC)[reply]

Hmm... Category:Articles for deletion should be updated every 24 hours; Category:Candidates for speedy deletion could be updated several times per hour (perhaps every 15 minutes), as every request is by default 200 pages, and would not be harsh on the servers. But overall, everything else looks fine. Titoxd(?!?) 07:37, 8 August 2006 (UTC)[reply]
One week, (up to 1 month if needed) trial period approved. — xaosflux Talk 01:27, 9 August 2006 (UTC)[reply]


This bot is a pywikipedia bot. It will be used semi-automatically for Interlanguage corrections. - Soulkeeper 12:21, 6 August 2006 (UTC)[reply]

One week trial approved, limit this to under 150 articles. Post difs/comments/etc here when trial is done. — xaosflux Talk 17:03, 6 August 2006 (UTC)[reply]
Diffs: Special:Contributions/Soulbot — I did make one or two mistakes in the beginning, but I've cleaned up my own mess, and now that I'm starting to learn the ropes of editing almost 40 wikipedias at the same time, the error ratio has dropped quite a bit. But I must also admit that the English Wikipedia is very updated interlanguage-wise, and it is probably the Wikipedia which has the least use for Soulbot. Most of its edits do not affect en. Soulkeeper 10:00, 13 August 2006 (UTC)[reply]
As you can see, most of the edits are: (adding no), but there are others too... ;) [23] [24] [25]- Soulkeeper 18:42, 17 August 2006 (UTC)[reply]

Please, give the bot flag to this bot. It is used for adding interwiki links with start on :cs:Wikipedista:JAnD. This bot has this flag on cs, is based on pywikipedia. All requested details are on it's userpage. Thanks, JAn Dudík 19:49, 29 July 2006 (UTC)[reply]

For reference, the userpage contains the following content: "This is unregistered manual controlled interwiki bot for years, days, cities and names (things, what are similar or mostly similar in different languages. Bot is controlled by [[:cs:Wikipedista:JAn_Dudík|JAn]]". I'm not really familiar with the approvals of interwiki linking bots, but you're not really giving us that much info about the bot. Can you provide some more clarification? alphaChimp laudare 20:08, 29 July 2006 (UTC)[reply]
Which details?
  • Bot's platform: pywikipedia, interwiki.py
  • Periodicity: always manual started on cs:, usually for one category or 10-20 pages.
  • There are many missing articles about czech towns on cs:, which have their article on en:. There is also many articles about years, where is on cs only link to en: but not backlink from en: or other languages.
  • My usual contribution on en is adding cs:. When do it I, why con't do it bot?
JAn Dudík 10:34, 30 July 2006 (UTC)[reply]
Your bot appears to be doing a lot of removals of interwiki links, what criteria is used for that? — xaosflux Talk 13:58, 30 July 2006 (UTC)[reply]
It seems to only be removing Japanese interwiki links. All of them are dead anyway. Maybe when it follows all the interwikis and automatically removes a link when it finds a blank page. Is that what it does Jan? alphaChimp laudare 14:07, 30 July 2006 (UTC)[reply]
Looks to be removing live links to, I don't read the other languages though, so can't tell what it is delinking. — xaosflux Talk 14:13, 30 July 2006 (UTC)[reply]
I couldn't find any removals of live ones. Everyone was seemed like the generic "Create a Page" wikipedia page (like you see here). Of course, I can't read Japanese, so I just assume the page meant what I thought it did. alphaChimp laudare 14:28, 30 July 2006 (UTC)[reply]
  • Please check these edits, and reply with the removal reasons:
  1. Edit: [26]
    Reason:
  2. Edit: [27]
    Reason:
  3. Edit: [28]
    Reason:
  4. Edit: [29]
    Reason:
Yes, you are right, when bot find dead link, automaticaly removes it.
ad1: on fr: is disambiguation, on other langs not
ad2: sl: do not exist, others are Town and villages (another category)
ad3: there are two gigues, the removed are'nt the same
ad4: there are two words: memetics and mem
all these changes were human assisted.
JAn Dudík 06:54, 31 July 2006 (UTC)[reply]
So, when and if this bot is running, will it be human assisted as it was during your test (in other words, will it be performing the type of removals Xaosflux was asking about above)? alphaChimp laudare 15:50, 31 July 2006 (UTC)[reply]
Interwiki bot can be running automatically, so when find two links for one page it will do nothing, when find non existing link he will add new links, in both cases write it to log on owners computer.
When is running with assisting, he asks every time when found more links to any languages which is the correct, and before removing link asks too.
JAn Dudík 11:27, 1 August 2006 (UTC)[reply]
Thanks for the replies, one week trial approved, please post difs and any comments here during run/when complete. — xaosflux Talk 00:30, 2 August 2006 (UTC)[reply]

Week is over, can you please unblock, rename ('JAnDbot) and flag this bot? Or may I wait for some more time? JAn Dudík 21:34, 9 August 2006 (UTC)[reply]

Looks like your trial was unable to be completed due to your block, I've left a message for the blocking admin to check on the status. — xaosflux Talk 00:01, 10 August 2006 (UTC)[reply]
Your bot has been unblocked, see the note on it's talk page. Trial period restarted. — xaosflux Talk 00:49, 11 August 2006 (UTC)[reply]

EssjayBot_III (talk · contribs · deleted contribs · nuke contribs · logs · filter log · block user · block log)

With the success of EssjayBot II archiving project and project talk pages, I've gotten several requests for it to do user talk pages. I'd prefer not to mix bot functions, just for my own peace of mind, and I didn't include user talk pages in my initial request, so I'm making a new request for this function. I personally am fine with doing it, or fine with not doing it; I'm just trying to provide those functions the people desire. Essjay (Talk) 10:08, 1 August 2006 (UTC)[reply]

Sure sure sure, trials approved, post some difs, etc...
How often will this script run, and how many pages will it deal with per interval? Having a bot archive hundreds of pages per interval could be a bit of a resource hog. This bot should definantley be flagged when approved. — xaosflux Talk 00:28, 2 August 2006 (UTC)[reply]
It'd run no more than once a day, and I'm generally very careful to make sure things run at times when other bots generally don't (for example, the archiving of ANI is done at 0:05 UTC, since most bots run on the hour). Since it's done by crontab entry, it can be done with whatever frequency; additionally, I can set them up load balanced if need be, putting three or four in one file, then three or four an hour later, etc. So far, I've had a handful of people ask; I have no idea how many Werdnabot was doing, but I really have no intention of getting into archiving the user talk pages of half of Wikipedia. Essjay (Talk) 03:17, 2 August 2006 (UTC)[reply]
Are you stil proposing this bot? (Will be removed in 1 week if no reply) — xaosflux Talk 00:26, 9 August 2006 (UTC)[reply]

Awaiting Operator Response


I would like to expand my AWB run bot to have it to subst user talk warnings in that namespace, per WP:Subst. I have been running it in for a while in manual, with no problems so far. Crazynas t 02:37, 31 July 2006 (UTC)[reply]

You do realize that we already have at least 4 bots (mine included) doing this, right? alphaChimp laudare 02:48, 31 July 2006 (UTC)[reply]
I also realize that when I loaded up my script in AWB there were around 60 pages to subst... if you have ideas for other things I can do, I'd be more then happy to listen. :). Crazynas t 02:55, 31 July 2006 (UTC)[reply]
Haha. Nah it's no problem. You can always check the Wikipedia:Bot requests page. Quite frankly, I don't think there's anything particularly wrong with having another bot subst'ing usertalk messages. Xaosflux - you agree? alphaChimp laudare 03:28, 31 July 2006 (UTC)[reply]
No problem having more of these, just makes for shorter runs, risk of bot collision is minimal at this. 1 week trial approved, post difs when complete. — xaosflux Talk 02:15, 1 August 2006 (UTC)[reply]

Awaiting Operator Response


BetacommandBot expansion of task

I would like to have the bot automatically tag catagorys older than 5 days with ((db-catempty}} that remain unused after five days Betacommand 05:20, 29 July 2006 (UTC)[reply]

How would it do that? What would the bot run in? How frequently would you run it? αChimp laudare 16:41, 29 July 2006 (UTC)[reply]

I will get a dump of all articles in Special:unused categorys and wait five days at that time i will get another dump and compare them manualy, any category that remains on the list is subject to deletion after only four days on the list. once i have the list of old empty cats i will then run the bot in AWB to tag each article with {{db-catempty}} per the deletion policy. I plan to run it no more than once a day, less as the number of empty cats goes down. currently there is 3926 empty catagories. Betacommand 17:26, 29 July 2006 (UTC)[reply]

An initial influx of 4000 CSD's would be a strain on anyone working CAT:CSD, can this be set to only flag new loney categories for CSD? — xaosflux Talk 22:46, 29 July 2006 (UTC)[reply]
Does unused cat's include categories that are "empty" of pages, but have subcats? — xaosflux Talk 22:46, 29 July 2006 (UTC)[reply]

I plan to use Special:Unusedcategories which states The following category pages exist although no other article or category make use of them. . as for the comment about flooding CSD i have a solution. create a category something to the effect of Category:Categories that have been empty for moer than five days and put a link to it on csd, so that the list can be accessed without flooding CSD and i will also maintain a record of the data I use at User:BetacommandBot/oldCategories — Preceding unsigned comment added by Betacommand (talkcontribs)

One comment about the current 3926 empties. I'm sure that a lot of them will remain empty, but is that the number of categories empty after 4 days, or is that just the total empty right now? (those two stats might actually be different) alphaChimp laudare 05:18, 30 July 2006 (UTC)[reply]
right now. update(3914) Betacommand 06:59, 30 July 2006 (UTC)[reply]
Using a holding category sounds good for the initial run, I've got no objection to using {{db-catempty}} tagging for the rest. Can you implement a check to prevent them from getting tagged on subsequent runs though (e.g. if in holding cat SKIP) ? — xaosflux Talk 03:14, 31 July 2006 (UTC)[reply]
already built in but thanks for the sugestion Betacommand 03:29, 31 July 2006 (UTC)[reply]
Would a trial run be possible without having to deal with the 3926 cats first? — xaosflux Talk 03:31, 31 July 2006 (UTC)[reply]

I plan to log them under User:BetacommandBot/oldCategories and slowly release the first large CSD over a period of time into the speedy. i was also thinking instead of {{db-catempty}} i create a similar template and put all Cats into it and place a link on WP:CSD so that they can handle the large number of cats as they get time without flooding CSD. Betacommand 03:45, 31 July 2006 (UTC) PS. kind of like a Backlog Category wich i will slowly release into CSD Betacommand 03:47, 31 July 2006 (UTC)[reply]

  • Trial period approved. Please limit the initial run to no more than 500 categories. If possible have the trial include some empty cats created after the trial starts. — xaosflux Talk 03:59, 31 July 2006 (UTC)[reply]
Trial run in progress please see User:BetacommandBot/oldCategories for full record and list of pages Betacommand 20:27, 1 August 2006 (UTC)[reply]
Can the per page logs check a master or prior page days to not regnerate so many hits per page? — xaosflux Talk 00:47, 2 August 2006 (UTC)[reply]
Trial run bloody well unapproved. He set it to tag for deletion hundreds and hundreds of categories that, whilst empty, are key elements in series of categories (like buildings by year, for years we don't have a buliding for yet). I've just sat and reverted them, and I don't have a bot. Much greater thought needs to be applied before approving expansions like this. Please; would this page stop handing approvals out for pointless tasks that don't need doing, and that need doing with a modicum of human judgement? Having witless bots do things merely for the sake of not having them idle is irritating at best. And this particular bot has already had one run of other edits (substs) reverted for also being wrong, and pointless. -Splash - tk 01:16, 2 August 2006 (UTC)[reply]
We do run short trials to see what may happen and require further review before just letting bots go. I'm not sure why we really need empty cats such as Category:5.56mm machine guns (one of the cats tagged). WP:CSD#1 does state that empty categories are speediable, and according to your statements above these may not even fall in to the old categories that may have contained articles before, requiring further investigation. If consensus is that we should have all of these empty categories, how did we end up with the speedy criteria? Regardless, due to this complaint, trials relating to editing the categories themselves are suspended, but ones that are preparing lists are not. Additionally, there has got to be a better way to keep track of the gathering then categorizing the categories. This bot has been blocked by Splash, pending a response. — xaosflux Talk 02:29, 2 August 2006 (UTC)[reply]
Well no, there were some that did seem a bit arbitrary, but I wasn't going to manually review each of 500ish categories when ~99% of them needed reverting. I do not think there is necessarily a "consensus" somewhere I can point to about the categories, but it does seem fairly obvious to me (and others at User talk:Kbdank71#Bot-tagging of unused categories) that obliterating large chunks of sequenced categories, just because they happen to be empty is wrong. Empty categories should be deleted when they risk confusion or duplication etc; that's what the CSD is for, not for indiscriminate application. Also, just because something meets a CSD does not mean it is required to be deleted, only that it may be.
Xaosflux, to answer your question, yes, I'm fine of course with someone unblocking the bot once its owner appreciates the difference in the manner it needs to be operated. Not merely asking here and getting an out-of-context nod before embarking on major-sized operations like this without checking very thoroughly, in advance that they make sense. WP:BOLD does not need to apply to bots. That said, I do not view the unblock as at all time-critical; the action of this bot fall well below the 'critical' threshold and in some cases well below the 'useful' threshold.
On a broader note, I'd like it if runs of 500 hundred edits were not the trial size. These have to be reverted by patient humans, and the whole point of a trial is to save time if things go belly up. I would think trials of no more than 50, with well announced trialling would be very much more appropriate. -Splash - tk 02:51, 2 August 2006 (UTC)[reply]
Responce:
  • A: My bot did not tag anything for deletion.
  • B: All it did was state that the category was subject to the old category rule WP:CSD#1.
  • C: If someone lets me know about a mistake i can have the bot revert all of the edits quickly without the overwork to humans.
  • D: It has been brought to my attention that some cats need to be kept. I would like assistance in creating a list of cats that are key elements in series of categories. i will use thwm as an exclusion
  • E: My bot did only 300-350 edits reguarding this subject
  • F: If at any point there is concern about my bot leave a message on its talkpage and it will stop editing until i have reviewed it and have resolved the issue

Betacommand 05:19, 2 August 2006 (UTC) PS: will work according to xaosflux Regardless, due to this complaint, trials relating to editing the categories themselves are suspended, but ones that are preparing lists are not. Betacommand 06:13, 2 August 2006 (UTC)[reply]

  • A, B: It amounted to tagging for speedy deletion.
  • E: That's still a factor of 10 too many for a trial run.
  • F: That is a useful feature that you should document on the bot's user page. -Splash - tk 19:33, 2 August 2006 (UTC)[reply]
What it is key to take away from this, Betacommand, is that you must check with relevant projects and processes and people before undertaking sweeping actions with a bot that are not precedented. Not before, to my knowledge have categories been mass eliminated under CSD C1, and so CfD should have been contacted before going ahead. This approvals page is evidently rather below-par for making sure of the utility and appropriateness of the editing the bot will undertake, really only making sure it is technically sound and possible. It is your responsibility to do the necessary homework. -Splash - tk 19:33, 2 August 2006 (UTC)[reply]
I think that this bot is a good idea. If someone where create the category, you expect it to have at least one item. Keeping them provides no navagationial value. --Shane (talk/contrib) 05:49, 2 August 2006 (UTC)[reply]
Well obviously sequential years have navigational value if I'm a newbie trying to work out what categories I can use without realising I can make them (particularly if I'm an anon and prohibited from creating pages at all). -Splash - tk 19:33, 2 August 2006 (UTC)[reply]
  • After receiving confirmation that this function is disabled for the time being (logs will still be written for debug purposes, but the bot will not make live edits related to this), I've unblocked the bot. However, if the bot is making any funny edits, feel free to reblock again. Titoxd(?!?) 06:07, 2 August 2006 (UTC)[reply]

I see no problem with this task. However, if certian categories are "required" or whatever, why not tag them with a template to that effect? Bots can detect the template and ignore it, and it will give humans the same cue. --Chris (talk) 02:13, 3 August 2006 (UTC)[reply]

I agree with Chris on this, a template would be a good idea, that makes the bot easily future-compatible, which is important for maintenance of series categories. I think that the task is a good idea, as long as some sort of exceptions mechanism is implemented, and all the needed categories are tagged before the bot runs.--digital_me(TalkContribs) 02:59, 3 August 2006 (UTC)[reply]

User:Splash I object to your to how you have handled this situation you are going against Wikipedia policy (empty categories older than four days are subject to deletion), and you inability to read.

This page meets Wikipedia's criteria for speedy deletion. It is a category that has no items whatsoever in it, it has been empty for at least four days (CSD C1).
Please Remove this notice if at any point this category is no longer empty


is how I marked the categories ‘‘‘NEVER’’’ did I list it for deletion or attempt to speedy it, all the bot did at this point was state 'This page meets Wikipedia's criteria for speedy deletion'. Regarding the fact that the categories are key elements in series of categories show some proof that there are guidelines to keep them and that they are exempt from WP:CSD#1. Also please show some consensus about keeping them that has more than four editors. That is NOT a consensus on Wikipedia. I am operating per Wikipedia:deletion policy. Please show me some guideline or policy that exist to back up your personal opinion, and the uncalled for hostile bordering on rude behavior you have show in this discussion. Betacommand 19:11, 3 August 2006 (UTC)[reply]

The CSDs are things that may be speedily deleted if an admin agrees with you. They are not things that must be. It does strike me that perhaps you don't agree that you should check with relevant projects, people and processes before deploying your bot. That's a little disappointing. -Splash - tk 23:44, 3 August 2006 (UTC)[reply]
I was planning to work with CSD to implament this task once i was completely comfortable with the logging and taging of old categories by the bot. you blocked my bot before i could do this. if you will look above i was thinking about putting a link on WP:CSD as a solution but i was planning to discuss this with CSD before nominating anything for deletion. As per above (my previous post) i am still waiting for the answers to the questions about your actions and the what policy that you were using for a guideline for reverting the tags and blocking of my bot. where is there a discussion about keeping them. and the other qusetions that i have rasied Betacommand 04:22, 4 August 2006 (UTC)[reply]

Ok, I am approving this bot on a trial run under the following conditions

  1. This trial is no more than 50 edits
  2. The bot will not tag w/ any tag that will add items to CAT:CSD, use a seperate sub-category like Tawkerbot on old user talk pages did.
  3. If it's a subcategory something should show to that extent (unless someone has a diff to prove otherwise) - if there is nothing there (say the category page is blank) for goodness sakes tag it!

-- Tawker 05:07, 4 August 2006 (UTC)[reply]

done, please see Category:Categories that have been empty for more than four days and are subject to deletion Betacommand 07:02, 4 August 2006 (UTC)[reply]
Nice…and you can see straight off the bat that those "A-class…" categories are each part of a set and should be tagged for exemption somehow. No mistake, this is a useful tool for rootling out categories which have gone unused and unloved, it just needs some mechanism for detecting those which are lying in wait for articles to populate them as part of an ongoing process.
One thing though: it would have been nice if the bot had spelt "Catagories" correctly HTH HAND —Phil | Talk 08:25, 4 August 2006 (UTC)[reply]
all A-Class articles are now excluced if there are any others that someone can identify i will add them to the exclude list thank you for your input. Betacommand 17:44, 4 August 2006 (UTC)[reply]

This is still no good. The trial has flagged a number of categories tagged as category redirects that have very obvious reason to exist for that purpose: to help prevent people categorising articles into them by seeing the red box that says "no, please put them in that category over there". It also tagged one category as meeting speedy criteria when it is currently on CfD. Betacommand, you seem to be unable to accept that BEFORE you do this, YOU NEED TO CHECK WITH SOME PAGE OTHER THAN THIS ONE. I can't make this any clearer than bold, italic capitals, and I've said as much several times earlier. The 'approvers' on this page do not apparently check that anybody actually cares whether or not your bot does stuff, and so you have to do that legwork yourself, before before before before before before before you jump in and do it. Now two of us have pointed out things wrong with this trial, I trust that it is unapproved - again. -Splash - tk 17:26, 7 August 2006 (UTC)[reply]

  • These kinds of things do not give me confidence at all. There's many categories that are empty, true, but that doesn't mean that they must be deleted. In fact, in most cases, they should not be, as they may be part of broader category schemes that aren't empty. There's no way to instill that kind of AI to a bot, so, as a result, I don't think a bot would be appropriate for this task. Oppose. Titoxd(?!?) 07:58, 8 August 2006 (UTC)[reply]
Responce: the bot has no AI I am its brain, I am currently working to identify all categories that should be kept like the A-Class artices you pointed out the bot only tags categories that i find, and even after they are identified as being empty for more that 4 days doent mean that they will be marked for deletion Because of some issuse that have been brought up durring this disscusion i was planning to manually examim all of the categories that meet the criteria befor putting them up for deletion. Just because my bot tags an article doesnt mean that it must be deleted or may deleted so far the tags only state that the category has been empty for more that four days and meeets the criteria for WP:CSD#1 it doesnt list them for deletion all it does is identify old empty categories. Given the issues raised durring this discussion I will review all categories thoroughly before i mark them for deletion. Betacommand 15:27, 8 August 2006 (UTC)[reply]

And if that does not seem to work for this task I have an alternate solution have the bot add all unused categories to a parent category such as category:Abandoned category as of ..... That Would allow users to quickly identify old empty categories by date. this would not involve marking categories for deletion or placing a template on the page it would be something like the orphaned category page only this Would identify abandoned categories Betacommand 05:41, 14 August 2006 (UTC)[reply]

More a question for the categories community, but perhaps a category "categories sometimes left deliberately empty" would help. Rich Farmbrough 21:03 15 August 2006 (GMT).
  • Given the issues involved with this Task i hearby withdraw my request Betacommand 16:04, 18 August 2006 (UTC)[reply]



  • What: orphaning (by linking in most cases) fair use images outside of ns:0
  • How: Based off lists generated by sql queries and reviewed to exclude some potentially ok pages (Portals there is some debate over, Wikipedia:Today's featured article and its subpages, Wikipedia:Recent additions, etc) using replace.py
  • How often: Probably in batches of ~500 pages based on the type of image (such as {{albumcover}}) list of images to remove list of pages to edit. With 10747 images and 19947 pages it may take a while still. Once this group is done, updates will depend on frequency of database dumps and/or whenever the toolserver works again and I can wrangle someone into running a report/get an account.

Kotepho 09:19, 12 July 2006 (UTC)[reply]

This one looks like it's going to be a magnet for complaints with people who don't understand image use policy but it does sound necessary. I'd start with a very well written our FAQ page and leave a talk page message on their page saying what the bot did and why it did it before I would run/approve it -- Tawker 21:36, 15 July 2006 (UTC)[reply]
Durin's page is quite good. Kotepho 21:49, 16 July 2006 (UTC)[reply]
Isn't this similar in at least some ways to what OrphanBot does? I'd like to hear Carnildo's comments on this, given that he runs OrphanBot and is on the approvals group. Essjay (Talk) 14:52, 16 July 2006 (UTC)[reply]
It is basically the same role, the main reason I brought it up is looking at OrphanBot's talk over time it does have a lot of complaints, (I suspect its people not familiar with policy mostly howerver) - it's just something FRAC, I have no problems with the bot personally and I'll give it the green light for a trial run, I just wanted to make sure Kotepho knew what deep water (complaints wise) this is :) -- Tawker 18:52, 16 July 2006 (UTC)[reply]
Well, someone has to do it, and I'd likely do at least some of them by hand so the complaints will come either way. Kotepho 21:49, 16 July 2006 (UTC)[reply]
Ahh, what the heck, run it in trial mode and lets see what happens -- Tawker 07:24, 21 July 2006 (UTC)[reply]

I did a run on the album covers, but skipped User: as they probably deserve a human touch (too many userboxes, page drafts, etc). Only had a few complaints, one was a typo in a regexp (since fixed)[30] that screwed up an infobox on a talk page and generally thinking archives should not be edited[31][32] and the other[33] was that removing the entire line of a gallery killed the caption[34], so I changed the behavior to just use a dummy image and prepend the imagelink to the caption[35] (caption was part of a discussion). Kotepho 21:19, 11 August 2006 (UTC)[reply]

Between February and April this year, I made a large number of typo-fixing edits (approximately 12,000 in total). All of these were done manually – every edit was checked before saving – although I have written software similar to AutoWikiBrowser to assist with the process. This software is designed specifically for spellchecking and so, while not as flexible as AWB, has a number of advantages. It reports the changes made in the edit summary, can check articles very quickly (in less than a second), and can easily switch between different corrections (for example, "ther" could be "there", "the" or "other") in a way that AWB cannot. Central to this is a list of over 5000 common errors that I have compiled from various sources, including our own list of common misspellings, the AutoCorrect function of Microsoft Office, other users' AWB settings, and various additions of my own. As I mentioned, I have done an extensive amount of editing with the aid of this software, using my main account. I have recently made further improvements to the software; over the last couple of days I have made a few edits to test these improvements, and I am now satisfied that everything works.

While I believe Wikipedia is now so heavily used that (a) no one person could hog the servers even if they wanted to, and (b) the Recent Changes page is more or less unusable anyway, a couple of users have expressed concerns about the speed of these edits (which reached 10 per minute during quiet periods). Most notably, Simetrical raised the issue during my RfA. As I stated in my response to his question, I was not making any spellchecking edits at that time, but I explained that I would request bot approval should I decide to make high-speed edits in the future. That time has now come; I have created User:GurchBot, and I request permission to resume exactly what I was doing in April, but under a separate account. I will leave the question of whether a bot flag is necessary to you; I am not concerned one way or the other.

Thanks – Gurch 19:45, 15 July 2006 (UTC)[reply]

As long as you are checking it yourself and ignoring the "sic"s, it seems good to me. Alphachimp talk 23:54, 15 July 2006 (UTC)[reply]
Yes, I check every edit before I save it, and I ignore [sic] when I see it. I have incorrectly fixed a couple of [sic]s in the past because I (the falliable human) failed to spot them; one of my improvements has been to add [sic]-detecting to the software so it can alert me to this, and hopefully make an error less likely in future – Gurch 10:03, 16 July 2006 (UTC)[reply]
  • I don't have any issue with this, provided you aren't doing any of the spelling corrections that tend to cause problems, such as changes from Commonwealth English to American English and visa versa. As long as it's only correcting spelling errors and doesn't touch spelling variations, it should be fine. I'd like to see a week's trial (which is standard) to get a good idea of exactly what will be taking place, and also for users to add their comments. A week's trial is approved, please report back this time next week. Essjay (Talk) 14:47, 16 July 2006 (UTC)[reply]
I have never corrected spelling variations, regional or otherwise – being from the UK, I have long since given up and accepted all variants as equally permissible anyway. If you wish, I can upload the entire list and replace the (now out-of-date) User:Gurch/Reports/Spelling; I will probably do this at some point anyway. I won't be around in a week's time, so you can expect to hear from me in a month or so. For now, you can take this to be representative of what I will be doing – Gurch 16:11, 16 July 2006 (UTC)[reply]
Just wanted to make sure I'd said it. ;) A month is fine; we normally do a week's trial, but I have no issues with something longer. Let us know how things are going this time next month. Essjay (Talk) 22:22, 16 July 2006 (UTC)[reply]

If these are manually-approved edits, I wouldn't think approval as a bot would be strictly necessary, though I could imagine the speed might be a concern, especially if errors are (or were) slipping through. Given that this is more of a "semi-bot", I suggest it not be bot-flagged, so as to reduce the likelihood of errors going undetected subsequently as well. Alai 04:24, 18 July 2006 (UTC)[reply]

In fact approval as a bot wasn't necessary – as I mentioned above, I used to do this using my main account, and would have continued to do so, except that a number of users expressed their concern and suggested I request approval for a bot. So I have done that. I freely admit that errors will inevitably slip through at some point; in fact, I've just had to apologize for correcting a British spelling which was, fortunately, spotted and reverted very quickly. Of course I never intended to do any such thing – it turns out that this (actually correct) spelling has been listed on Wikipedia:Lists of common misspellings/For machines (one of the sources for my correction list) since November 2002; somehow it was never spotted in nearly four years. My fault, of course, for assuming the list was correct; I'm now scrutinizing my list thoroughly to avoid repeating this mishap. This is the first time I've made such a miscorrection, the reason being that my old list was constructed by hand, whereas I've now tried to expand it (and so catch more errors with each edit) by including lists from other sources. In the past I have occasionally mis-corrected "sic"s and errors in direct quotations; the chance of this should be much lower now that my software can detect these itself, even if I miss them. Based on what I have done to date, though, I reckon my error rate is about 1 in every 1000 edits, which I can live with – Gurch 11:38, 18 July 2006 (UTC)[reply]
  • As I said above, you're cleared for a month-long (instead of a week, at your request) trial; check back with us then and we'll set the bot flag. Essjay (Talk) 00:47, 19 July 2006 (UTC)[reply]

Concern The list of common mispellings is utter shit, please do NOT use it. It replaces many words that are actually words. --mboverload@ 20:34, 28 July 2006 (UTC)[reply]

I'm not (any more). I've found some more reliable lists from other sources. I would use your list, but it would take forever to strip the regexes out. (I did think of implementing regex support in my software, but I think it would slow it down too much) – Gurch 18:47, 14 August 2006 (UTC)[reply]
The overhead of regexing is modest compared with d/ling the page, uploading the changed text, d/l the diffs, uploading the final changes and d/l the final page. Rich Farmbrough 21:05 15 August 2006 (GMT).

Proposed disambiguation bot, manually assisted, running m:Solve_disambiguation.py. I will be using this to work on the backlog at Wikipedia:Disambiguation pages with links; bot assisted disambiguation is substantially more efficient than any other method. The bot will run from the time it gets approval into the foreseeable future. --RobthTalkCleanup? 16:20, 13 June 2006 (UTC)[reply]

I see no reason we couldn't have a trial run, at least. robchurch | talk 20:44, 1 July 2006 (UTC)[reply]
Thanks. I'll start running it at low speed in the next couple of days. --RobthTalk 04:04, 2 July 2006 (UTC)[reply]
(In response to a request for a progress report): I've made a small run, which went quite well, but limits on my disposable time have prevented me from making any larger runs just yet--RobthTalk 01:07, 10 July 2006 (UTC)[reply]
  • No problem, trial extended, keep us informed and report back when you have enough done for us to make a final decision. Essjay (TalkConnect) 08:24, 12 July 2006 (UTC)[reply]

Awaiting Operator Response

Approved




Approved, needs to be flagged


Approved, flagged



Denied/Withdrawn





Wikipedia:Bots/Requests for approvals/Historybot


Expired Requests

These requests have expired, as information required by the operator was not provided. These bots are NOT authorized to run, but not on the basis of merit. An example of a bot for this section would be a bot that was approved for testing, but then not tested, or did not have posted testing results. Bot requests should not be placed here if there is an active discussion ongoing above. Operators who's requests have expired may reactivate their request at anytime.