Jump to content

User talk:Beetstra: Difference between revisions

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
Content deleted Content added
SineBot (talk | contribs)
m Signing comment by Jamesstatham - "→‎Dan Ariely: new section"
Line 516: Line 516:


Thanks <small><span class="autosigned">— Preceding [[Wikipedia:Signatures|unsigned]] comment added by [[User:Jamesstatham|Jamesstatham]] ([[User talk:Jamesstatham|talk]] • [[Special:Contributions/Jamesstatham|contribs]]) 17:26, 21 February 2012 (UTC)</span></small><!-- Template:Unsigned --> <!--Autosigned by SineBot-->
Thanks <small><span class="autosigned">— Preceding [[Wikipedia:Signatures|unsigned]] comment added by [[User:Jamesstatham|Jamesstatham]] ([[User talk:Jamesstatham|talk]] • [[Special:Contributions/Jamesstatham|contribs]]) 17:26, 21 February 2012 (UTC)</span></small><!-- Template:Unsigned --> <!--Autosigned by SineBot-->

:Thank you for your question. I'll try and explain. First of all, we are writing an encyclopedia here, not a linkfarm or an internet directory. Pages should be as much as possible self-contained. However, Wikipedia can not incorporate all information. Therefore, we can incorporate external links. However, it lies not in Wikipedia's goal, to link to every resource about a subject. If I see the page you were linking to, I do not see a lot of information that is not already in the Wikipedia page, moreover, most of it could be incorporated easily. Then, you use as a description of the link "Represented for speaking engagements by Leigh Bureau" - that looks certainly promotional. Were you linking to the site to give more information about the subject (which I don't think the site has), or were you linking to promote (I do not regarding the latter, that you [http://en.wikipedia.org/w/index.php?title=John_Micklethwait&diff=prev&oldid=478056983 first change an existing link], [http://en.wikipedia.org/w/index.php?title=John_Micklethwait&diff=next&oldid=478056983 then change the text], and [http://en.wikipedia.org/w/index.php?title=John_Micklethwait&diff=next&oldid=478085634 then move the link higher up so it looks more prominent]. I would really suggest that you heed the warning that Dmacks left you. --[[User:Beetstra|Dirk Beetstra]] <sup>[[User_Talk:Beetstra|<span style="color:#0000FF;">T</span>]] [[Special:Contributions/Beetstra|<span style="color:#0000FF;">C</span>]]</sup> 18:31, 21 February 2012 (UTC)

Revision as of 18:31, 21 February 2012

Welcome to my talk page.

Please leave me a note by starting a new subject here
and please don't forget to sign your post

You may want to have a look at the subjects
in the header of this talkpage before starting a new subject.
The question you may have may already have been answered there

Dirk Beetstra        
I am the main operator of User:COIBot. If you feel that your name is wrongly on the COI reports list because of an unfortunate overlap between your username and a certain link or text, please ask for whitelisting by starting a new subject on my talkpage. For a better answer please include some specific 'diffs' of your edits (you can copy the link from the report page). If you want a quicker response, make your case at WT:WPSPAM or WP:COIN.
COIBot - Talk to COIBot - listings - Link reports - User reports - Page reports
Responding

I will respond to talk messages where they started, trying to keep discussions in one place (you may want to watch this page for some time after adding a question). Otherwise I will clearly state where the discussion will be moved/copied to. Though, with the large number of pages I am watching, it may be wise to contact me here as well if you need a swift response. If I forget to answer, poke me.

I preserve the right not to answer to non-civil remarks, or subjects which are covered in this talk-header.

ON EXTERNAL LINK REMOVAL

There are several discussions about my link removal here, and in my archives. If you want to contact me about my view of this policy, please read and understand WP:NOT, WP:EL, WP:SPAM and WP:A, and read the discussions on my talkpage or in my archives first.

My view in a nutshell:
External links are not meant to tunnel people away from the wikipedia.

Hence, I will remove external links on pages where I think they do not add to the page (per WP:NOT#REPOSITORY and WP:EL), or when they are added in a way that wikipedia defines as spam (understand that wikipedia defines spam as: '... wide-scale external link spamming ...', even if the link is appropriate; also read this). This may mean that I remove links, while similar links are already there or which are there already for a long time. Still, the question is not whether your link should be there, the question may be whether those other links should be there (again, see the wording of the policies and guidelines).

Please consider the alternatives before re-adding the link:

  • If the link contains information, use the information to add content to the article, and use the link as a reference (content is not 'see here for more information').
  • Add an appropriate linkfarm like {{dmoz}} (you can consider to remove other links covered in the dmoz).
  • Incorporate the information into one of the sister projects.
  • Add the link to other mediawiki projects aimed at advertiseing (see e.g. this)

If the linkspam of a certain link perseveres, I will not hesitate to report it to the wikiproject spam for blacklisting (even if the link would be appropriate for wikipedia). It may be wise to consider the alternatives before things get to that point.

The answer in a nutshell
Please consider if the link you want to add complies with the policies and guidelines.

If you have other questions, or still have questions on my view of the external link policy, disagree with me, or think I made a mistake in removing a link you added, please poke me by starting a new subject on my talk-page. If you absolutely want an answer, you can try to poke the people at WT:EL or WT:WPSPAM on your specific case. Also, regarding link, I can be contacted on IRC, channel [1].

Reliable sources

I convert inline URL's into references and convert referencing styles to a consistent format. My preferred style is the style provided by cite.php (<ref> and <references/>). When other mechanisms are mainly (but not consistently) used (e.g. {{ref}}/{{note}}/{{cite}}-templates) I will assess whether referencing would benefit from the cite.php-style. Feel free to revert these edits when I am wrong.

Converting inline URLs in references may result in data being retrieved from unreliable sources. In these cases, the link may have been removed, and replaced by a {{cn}}. If you feel that the page should be used as a reference (complying with wp:rs!!), please discuss that on the talkpage of the page, or poke me by starting a new subject on my talk-page

Note: I am working with some other developers on mediawiki to expand the possibilities of cite.php, our attempts can be followed here and here. If you like these features and want them enabled, please vote for these bugs.

Stub/Importance/Notability/Expand/Expert

I am in general against deletion, except when the page really gives misinformation, is clear spam or copyvio. Otherwise, these pages may need to be expanded or rewritten. For very short articles there are the different {{stub}} marks, which clearly state that the article is to be expanded. For articles that do not state why they are notable, I will add either {{importance}} or {{notability}}. In my view there is a distinct difference between these two templates, while articles carrying one of these templates may not be notable, the first template does say the article is probably notable enough, but the contents does not state that (yet). The latter provides a clear concern that the article is not notable, and should probably be {{prod}}ed or {{AfD}}ed. Removing importance-tags does not take away the backlog, it only hides from attention, deleting pages does not make the database smaller. If you contest the notability/importance of an article, please consider adding an {{expert-subject}} tag, or raise the subject on an appropriate wikiproject. Remember, there are many, many pages on the wikipedia, many need attention, so maybe we have to live with a backlog.

Having said this, I generally delete the {{expand}}-template on sight. The template is in most cases superfluous, expansion is intrinsic to the wikipedia (for stubs, expansion is already mentioned in that template).

Warning to Vandals: This user is armed with VandalProof.
Warning to Spammers: This user is armed with Spamda
This user knows where IRC hides the cookies, and knows how to feed them to AntiSpamBot.

ChEMBL Multiple IDs

Hi Dirk, could you please program the DrugBox to accept 2 ChEMBL IDs so that I can add parents and salts? The ChemBox works like this already and it's excellent. Thanks in advance, Louisa Louisajb (talk) 11:57, 9 December 2011 (UTC)[reply]

I am sorry, I really do not have time at the moment. Would you mind posting on Template talk:Drugbox, maybe one of the other regulars of the drugbox is willing to help. Have a good time, and I expect that it will be next year before we see again (I am really busy). Thanks! --Dirk Beetstra T C 11:52, 19 December 2011 (UTC)[reply]

Template:Chembox subst explosive has been nominated for deletion. You are invited to comment on the discussion at the template's entry on the Templates for discussion page. Bulwersator (talk) 14:46, 9 December 2011 (UTC)[reply]

I see this has been solved. All those templates are for some reason miscategorised (and I must confess, I have no clue if and how many people use this template ..). Thanks! --Dirk Beetstra T C 11:53, 19 December 2011 (UTC)[reply]

WP:UWTEST update

Hi Beetstra,

We're currently busy designing some new tests, and we need your feedback/input!

  1. ImageTaggingBot - a bot that warns users who upload images but don't provide adequate source or license information (drafts here)
  2. CorenSearchBot - a bot that warns users who copy-paste text from external websites or other Wikipedia articles (drafts here)

We also have a proposal to test new "accepted," "declined," and "on-hold" templates at Articles for Creation (drafts here). The discussion isn't closed yet, so please weigh in if you're interested.

Thanks for your help! Maryana (WMF) (talk) 02:07, 15 December 2011 (UTC)[reply]

Sorry, I have no time, though I would like to interact about this. I hope I have more time again next year. Thanks for the notice, see you around! --Dirk Beetstra T C 11:54, 19 December 2011 (UTC)[reply]

Ending XLinkBot test

Hey Beetstra, just a heads up that Maryana and I planned on ending the XLinkBot template test in a day or two, since it was started on the 17th of last month. I just wanted to double check: is this a revision you wanted to keep regardless of the template contents? Hope you're well, Steven Walling (WMF) • talk 19:55, 15 December 2011 (UTC)[reply]

Yep, that needs to stay (or re-added, which is maybe easier) - those parameters tell m:User:LiWa3 and User:XLinkBot where the revertlists are, the bots accept multiple, and some trusted non-admin users (those who hence do not have access to the main list), and whom I trust enough to use the bots have access to private lists via these settings.
I am away for a lot of time, and may not have time to go onto Wikipedia until somewhere in the beginning of next year (and even then). I will have to leave bot-operation to Versageek (who has access to all my bots), maybe further questions can be redirected to Versageek (but please keep me posted here, as I'd like to know)? Thanks! --Dirk Beetstra T C 11:58, 19 December 2011 (UTC)[reply]
Okay, rollback of the setting (plus your recent edit to the revertlist) is  Done. We'll keep you and Versageek updated on analysis work, since we're starting a pretty intensive round of it. Thanks for everything Beetstra, and happy holidays. Steven Walling (WMF) • talk 18:42, 19 December 2011 (UTC)[reply]

ChemBox data enrichment

Hi again Beetstra, I posted previously about a 670 drug database with human PK data. Probably I inserted it wrong, sorry about that.

Where you replied that it was a problem to read in the CAS identifiers, since you use InChi. I managed to convert the list to InChi. I can forward it to you... BUT how? Wouldnt wanna litter your talk page.

Just for completeness sake the list is found here http://dmd.aspetjournals.org/content/suppl/2008/04/21/dmd.108.020479.DC1/DMD20479_Trend_analysis.xls

Warm greetings ./Claus — Preceding unsigned comment added by 109.118.41.138 (talk) 15:47, 17 December 2011 (UTC)[reply]

I may not have access to do this work again until somewhere in the beginning of next year. I will try to remember and have a look then. Thanks! --Dirk Beetstra T C 11:59, 19 December 2011 (UTC)[reply]

Psilocybin CAS number

Hi Dirk. I hope you are enjoying the holidays. When you get the chance, would you mind taking a look at psilocybin and see why CheMoBot is not verifying the CAS number even though I listed it at User:Edgar181/non-commonchemistry-sourced CASNo. User:Sasata is trying to get the article to FAC and is taking a close look at everything there. Thank you. -- Ed (Edgar181) 02:37, 28 December 2011 (UTC)[reply]

I did not take it over yet, and CheMoBot does not read that list. All what CheMoBot does goes through the index (Wikipedia:WikiProject Chemicals/Index/Wikipedia:WikiProject Pharmacology/Index). I'd suggest that all the relevant identifiers are properly checked, and that a revid of the page where all those are either correct or blank (blank for those for which do not exist or which are not verifiable to one correct value) is added to the correct index. I hope this explans. --Dirk Beetstra T C 17:35, 8 January 2012 (UTC)[reply]
I think I understand. I'll see what I can do. Thanks. -- Ed (Edgar181) 14:05, 9 January 2012 (UTC)[reply]

pihasurfschool.com

Hi there We we blacklisted. We try to add an external link to A History of Surfing in Piha www.pihasurfschool.com/about-piha.html to wiki page about Piha. We live in Piha and we are surfers. Can you please allow our link? It not a commercial page. Regards, Phil. — Preceding unsigned comment added by 125.236.243.225 (talk) 10:06, 3 January 2012 (UTC)[reply]

Beetstra is on a wikibreak so I will provide some suggestions.
This relates to Piha. Please review WP:EL and consider whether the external link you propose would actually assist readers of the article. If you have good reasons for why the link would be helpful, please post a request at WT:WHITELIST.
Are you sure the link is blacklisted? Or was it simply reverted by other editors? Johnuniq (talk) 11:34, 3 January 2012 (UTC)[reply]
As far as I can tell, the link has not (yet) been blacklisted. I suggest you discuss placing the link before adding it to any article on any language. I think this link does not comply with the guidelines for external links and should not be in wikipedia. EdBever (talk) 19:31, 3 January 2012 (UTC)[reply]
I've started a discussion of the link at Talk:Piha. Please contribute there.-gadfium 20:03, 3 January 2012 (UTC)[reply]

I answered there, IMHO, the link does not belong there per WP:NOT#REPOSITORY and WP:NOT#DIRECTORY and the intro of WP:EL. --Dirk Beetstra T C 17:33, 8 January 2012 (UTC)[reply]

Publications Office of the European Union / Publications Office pages

Dear Mr Beestra,

I am the social media editor for the Publications Office. We understand that Facebook has generated pages from Wikipedia called 'Publications Office of the European Union' and 'Publications Office'.

Our first concern is that we have created a page on Facebook 'EU Law and Publications'in which we feed our latest news. But we think that people are being confused with the automatically generated pages. I would therefore like to redirect people to our bona fide Facebook page. This is why I have tried to change the pages.

Can you help me please? — Preceding unsigned comment added by 158.169.9.14 (talk) 11:30, 12 January 2012 (UTC)[reply]

Thank you for your question. That facebook page fails our inclusion standards, and that is why I have removed them all (and also XLinkBot is trying to tell you that). We do not have to include all official pages or everything that is related to the subject. Please read the external links guideline and 'What Wikipedia is not'. Thank you. --Dirk Beetstra T C 14:45, 12 January 2012 (UTC)[reply]

Chembox property count updated

See User:Itub/Chembox property count. Sorry it took so long to reply! --Itub (talk) 21:58, 13 January 2012 (UTC)[reply]

Thanks Itub. Good to see that the number of parameters is quickly increasing. I also see many 'broken' parameters. I'll try to work on that one of these days. --Dirk Beetstra T C 04:06, 17 January 2012 (UTC)[reply]

A barnstar for you!

The Technical Barnstar
You did the migration of COIBot, and as you must be beer free, please have this very expensive award and the beer will have to be taken on notice. — billinghurst sDrewth 03:40, 17 January 2012 (UTC)[reply]
Thanks! I'll wait for better beer times in stead of having one of the alcohol free beers they sell here in the malls. I don't want to share the faith of the Buckler drinkers in the Netherlands. --Dirk Beetstra T C 04:05, 17 January 2012 (UTC)[reply]
Oh, I was thinking more a tasty drink. — billinghurst sDrewth 06:29, 19 January 2012 (UTC)[reply]
Oh, that for sure, billinghurst. But the main beer I saw here in the supermarkets (only visited one huge one so far) is, low and behold, alcohol free Budweiser (imagine being in dry country, and seeing from the other side of the shop a stand with bottles of which you can only make up the word 'Budweiser' from that distance). They do have a couple of other brands of beer (many are of the fruity type, like the Belgian Kriek), but it is quite minimal (and of course all is alcohol free). I never tried James Squire, I am looking forward to it! --Dirk Beetstra T C 07:23, 19 January 2012 (UTC)[reply]

CoiBot question

Hi Dirk. I've left a comment at Wikipedia:WikiProject Spam/Local/apps.gov.bc.ca. I was asked to look into it by new user User:Msruzicka, who was concerned that he was in trouble. I had actually told him to add BCGNIS info to some of his many BC geo stubs to flesh them out, some I'm partially responsible. I haven't worked with one of these reports before, so maybe my comment is improperly placed. If you have time, maybe you could give it a look. Thanks, The Interior (Talk) 06:00, 19 January 2012 (UTC)[reply]

All is sweet. It is monitoring the link, not reporting it as spam. The text in the lead box is relevant to set the scene, and that the relevant site is not entered in any of the Links list is the relevant concern. As it has been reviewed (by me), I have closed it as such. If another batch of links are added, then it will probably open again, and in which case we can close it again. No issue, it is just doing its job of monitoring. — billinghurst sDrewth 06:26, 19 January 2012 (UTC)[reply]
Thanks, billinghurst! The Interior (Talk) 07:05, 19 January 2012 (UTC)[reply]

(ec) Thanks, The Interior and billinghurst. I will expand a bit on this.

Indeed, the bot is doing its job monitoring. The location of these reports is with the WikiProject Spam, as they are the main ones monitoring these reports. The name of the project (and the WikiMedia feature 'MediaWiki:Spam-blacklist') is a bit misleading. Quite a bit of the material handled by the project and quite a number of the entries on that list are not 'spam' even under a wide definition of that term.

We run a number of programs which monitor all link-additions by all editors to Wikipedia (actually, 772 wikimedia projects), with as a main goal to catch spam or other inappropriate stuff. All those link additions are stored in a big database. Now, real spam (on Wikipedia widely defined as 'links added for promotion', which includes what the general public thinks of as spam, like porn, viagra etc., but it actually also includes regular companies optimizing their search results (Search Engine Optimization, SEO), public organisations who want to make their (generally good) cause known to the public to raise more money and politicians wanting to increase people finding their pages so they would gain more votes) generally has as a common feature that they are 'new links added by a new user' (those links are never used before, so they do not appear in the database, and those editors often did not edit before). So if we find a new user focusing on one link then that is reason for concern. The bot notifies the Wikipedia editors by automatically creating a linkreport. (the catching system is stronger than that, but I will of course not disclose all the exact features per WP:BEANS)

A good percentage of those links are added in good faith (although that does not make it right all the time) and can be ignored (or in some cases, just reverted and then ignored), the rest is real spam and further action can be taken then. As billinghurst already noticed, these links are good, and in a way, we just ignore the link additions. I reporting persists, I would suggest to whitelist it hard (which needs to be done off wiki) or at least set the status of the report to 'ignore'. Note that the reports are not indexed by search engines (that follow that directive, the major ones do), they can not be found on internet, you have to specifically go to Wikipedia to find them.

I hope this explains. --Dirk Beetstra T C 07:17, 19 January 2012 (UTC)[reply]

It certainly does, appreciate it Dirk. As the majority of appropriate pages have the link already, it probably won't come up again. Now that I've read the primer for fighting spam, let me know if you guys need help with any backlogs, etc. Best, The Interior (Talk) 00:38, 20 January 2012 (UTC)[reply]

Update: new user warning test results available

Hi WP:UWTEST member, we wanted to share a quick update on the status of the project. Here's the skinny:

  1. We're happy to say we have a new round of testing results available! Since there are tests on several Wikipedias, we're collecting all results at the project page on Meta. We've also now got some help from Wikimedia Foundation data analyst Ryan Faulkner, and should have more test results in the coming weeks.
  2. Last but not least, check out the four tests currently running at the documentation page.

Thanks for your interest, and don't hesitate to drop by the talk page if you have a suggestion or question. Maryana (WMF) (talk) 19:23, 20 January 2012 (UTC)[reply]

Thanks, that is interesting. I'll have a look. In case I forget, one thing I was pondering the last couple of weeks sometimes about this test (not sure how to put this into words): I had the feeling that the reasoning for complaining to XLinkBot or reverting was different during the test - the type of complaints felt different. I know that the current warnings are long, but I have the feeling that the complainers show more understanding now than during the test (you have to separate the complaints about the mode of operation of the bot from the complaints about the links itself). Note that there are (outside of the testing) editors who come to XLinkBot with 'I've read WP:EL, and I think you are right, the link was not suitable, I have adapted the edit'. I don't know if there is enough data recorded to really get hard data about that. It may be reflected in the number of re-additions after the bot removes - I am afraid that the test-templates did not let the editors understand why the link was removed, and that they therefore blindly revert the bot (and if they complain, they give me the other feeling about the complaints, showing that they did not understand why they were reverted by the bot in the first place). --Dirk Beetstra T C 19:56, 20 January 2012 (UTC)[reply]
I think your experience with that is correct. This is likely a case of language that works really well with one kind of new editor (such as test edits, like adding [[File:Example.jpg]]) does not work well with a different kind of editor. It sounds entirely reasonable to me that people adding really wrong external links need more education.
I do want to say though, that in combing through the test data and in my personal editing, I sometimes see someone who made other valuable text additions get reverted because one part of their edit was to add an inappropriate link. Those editors are usually confused and think it means all their contributions were bad, not just the link. Perhaps one way forward is to keep the current non-test versions of the templates, but make XLinkBot much more cautious about reverting someone who added several bytes of text along with a bad link -- maybe it could just log those edits and warn the editor? Steven Walling (WMF) • talk 21:13, 20 January 2012 (UTC)[reply]
I don't think that would be desirable. Wikipedia is a sitting duck for external link promotions (covered by XLinkBot) and vandalism (covered by ClueBot), and IMHO the most important remedy is speed of response. People adding external links count any exposure as a win—if an external link is visible for a few hours before a human removes it, the person promoting the external link will often think their time was well spent and they got a good result. What discourages external link spammers and vandals is when a bot reverts their edits within minutes. Of course the situation is not clear, and XLinkBot reverts good-faith (although possibly naive) editors as well as others of less good faith. There is no good solution, and the current system may discourage a small number of editors who start by adding several external links. However, many people whose primary aim is to add external links and who manage to maintain their links for an extended time, then develop a sense of entitlement and require extensive volunteer effort when experienced editors try to remove external links per WP:EL. There would need to be evidence of a real problem (the loss of potentially good editors) to warrant changing the behavior of the bot, and one important issue is that someone who cannot read and understand a reasonable message is not showing promise. Johnuniq (talk) 23:19, 20 January 2012 (UTC)[reply]
Johnuniq: I should say that I did a quick bit of qualitative coding on the users who were warned (you can see it for yourself in the XLinkBot data spreadsheet here), and the majority (79 out of 100) were not spammers – they were simply linking to YouTube, Facebook, Twitter, or Wordpress. I'd be happy to do some more coding on this sample, and you're welcome to, as well (just ask and I'll give you editing privileges), but I'm guessing that it's pretty representative of who XLinkBot is hitting. So while serial spammers who crave exposure do exist, they're actually pretty rare. Maryana (WMF) (talk) 00:07, 21 January 2012 (UTC)[reply]

I took the first 12 cases that were not shown as "spam" from the above mentioned spreadsheet, and checked the edits, with these results:

There is nothing in the above results to suggest that changing XLinkBot would be helpful. It is good that the WMF wants to encourage editing, but there appears to be a belief that the community has an inexhaustible supply of good editors who are available and willing to dedicate an hour to explaining the purpose of Wikipedia to new arrivals. Judging from the above, the bot has saved a great deal of time and trouble. If there is a problem, one approach may be to encourage a group of editors to follow the bot's work, and to revert the bot (and remove its warnings) where appropriate, and to manually engage with the new editor. Johnuniq (talk) 03:19, 21 January 2012 (UTC)[reply]

I see you make the crude separation between youtube/facebook/myspace/twitter etc. and 'spam'. There are a couple of major concerns there which one needs to take into account.

  • A part of youtube/facebook/myspace/twitter etc. are spam nonetheless. They are added for a promotional target. Above, those who create/change articles to bio's of musicians with facebook links are not naively adding facebook links, quite some of them are here to promote.
  • Youtube.com - I did a quick scan of 10 reverts a couple of months ago. Two of those (20%!) were links to video clips of songs of artists which were still within copyright, and not obviously uploaded by the owner of those rights - in other words, very likely they were copyright violations. See the blackout of a couple of days ago, I think they are an even bigger concern than the spam part. Do note that YouTube is not the only site that hosts copyvios, that is also true for blogspot e.g. (a long time ago I had an admin complaining to me that I reverted a blogspot reference, he did not see anything wrong with it - it was however a straight copy of a newspaper article where the article was behind a (albeit free) login, the article belongs to the newspaper, the blogspot was a copyvio of it).
  • Also note that wordpress/blogspot and other blog-like sites are also often added promotional by the writer of the piece. It may sometimes be helpful, nonetheless it qualifies to something close to spam.

I think that the bot explains now properly that if the edit contained more than only an offending link, that the editor should undo with the consideration to remove the external link before saving the page again. Also that was something that was not said in the test-warnings, and editors may therefore have been more inclined to revert the whole edit without consideration ...

I'll answer later more, when I have more time. --Dirk Beetstra T C 04:17, 21 January 2012 (UTC)[reply]

Following the conversation a little from the outside, and more focusing comment from the external view of xwiki abuse rather than internal abuse, and not had the time to do any analysis on the data or the heuristics of Xlinkbot; though I will qualify that statement with I have done a lot of cleanup work recently in m:Category:Open Local reports for en.wikipedia.org. Help:Citing sources pretty clearly considers blogs and many of the mentioned as less authority as references. So that gives scope to the sort of message that we can deliver, reverted or warned. There is the issue of external links, versus <ref> links, and the placement of the link. To my anecdotal feedback, the higher it is placed in an existing list, the more problematic. Similarly, I see numbers of issues where people add the section External links and add a link, or the only pre-existing link is a template and the following link is promotional. Clearly the less text and more link that is added is an indicator of an issue, similarly lack of addition of refs, though both can also reflect newbieness and nervousness/tentativeness in an edit. It is the balance between adding quality references, and exceptional external links to build a quality encyclopaedia, and finding the the nice way to do it when it needs to be done in circumstances of good faith. — billinghurst sDrewth 05:33, 21 January 2012 (UTC)[reply]
Addendum. From the results available it is not evident to me that I can easily see where the focus is on the message provided or the type of reverted edit. Plus from the xwiki work, I am less likely to be be seeing xlinkbot in action, but more what it didn't catch, which may skew things in a positive or a negative direction. — billinghurst sDrewth 05:39, 21 January 2012 (UTC)[reply]

Just to expand on the point by Johnuniq regarding separating 'spam' from 'youtube/blogs/etc', and in reply to Maryana (WMF):

Do take care. You say 'the majority (79 out of 100) were not spammers – they were simply linking to YouTube, Facebook, Twitter, or Wordpress' .. two of these sites here now were clearly spammed (both reported by XLinkBot to be blocked for spamming, both blocked as a result of that). --Dirk Beetstra T C 19:22, 22 January 2012 (UTC)[reply]

Johnuniq's original point, if I understand it correctly, was that people who insert inappropriate external links into articles are mostly doing so to gain revenue from directing clicks and eyeballs to their own sites. My response was that YouTube, Flickr, and Facebook/Twitter linkers, who make up a significant chunk of people warned by XLinkBot, are qualitatively different from spammers who insert links to obscure commercial or promotional websites. They're not linking to copyvio videos or fan pages because they want ad revenue – they're linking to them because they don't understand the rules of Wikipedia and think their contribution is acceptable. If they don't get the message after repeated warnings and blocks, it's not because they're hopelessly stupid; it's because we're not doing a good enough job of explaining the rules to them in a clear, non-hostile way. I understand that it gets frustrating having to explain the same thing to people day in and day out, but you have to remember that what feels so strikingly obvious to you, an experienced Wikipedian, is still incredibly strange and incomprehensible to new users.
As for blogs, for every case of self-promotion, there are cases like this, where someone tries in good faith to add a reference he thinks is appropriate, and the only help or feedback he receives comes in the form of a template warning from a bot. That's why I'm saying we should pay more attention to these messages, and why we should continue to tweak them until they're at least as helpful as a human mentor. I do understand that Wikipedia doesn't have an army of people to patiently guide newbies through stuff like this. That's exactly why I think it's so important to get template messages right. But it's a chicken-and-egg dilemma. Part of the reason we don't have that army is that our editing numbers are dwindling by the day, and part of the reason for that is that we don't do enough to help and guide new users. Maryana (WMF) (talk) 19:04, 23 January 2012 (UTC)[reply]

I agree, Maryana, Wikipedia there has a problem, retaining editors, gaining new editors. Notifying people appropriately and properly is there necessary. That does not necessarily lie in a short message without flooding them with policies and guidelines, it lies in bringing a nice and friendly message notifying the people appropriately of the point you want to get through. A too short a message, however friendly, does not get the message through, and people will react more with 'go away bot, I think my link is good', they revert, and don't get to know why some parts are a problem.

I understand that you meant, but the qualification that editors who add twitters, youtubes and the like are mainly doing that not for gaining revenue but because of ignorance of the rules is too short. The obscure website owners are indeed the ones who do it for revenue (not necessarily obscure websites, we've had large organisations of continental or global importance spamming Wikipedia for revenue), but a significant part of the twitters, youtubes and myspaces are not added because of ignorance, they are added solely for promotion.

And that is where we hit the problem, I think. Many people regard spam as a form of vandalism, the editor adding the inappropriate links for any form of promotional effort is just an editor who is ignorant of Wikipedia's rules. Sure, some of the editors who add a large number of inappropriate links do it with the idea that they think they are helping Wikipedia, but the true spammers are not here for that reason. And that separates spammers from the good faith editor who is ignorant of our rules. Most of the editors out of the second group only find one, very maybe two, warnings by XLinkBot. That initial message there is aimed at being friendly, and gaining understanding. Only very rarely I encounter good faith editors who get to a {{uw-spam4}} (I recall out of the last year two cases, one genuine editor whose editing was a bit clumsy and who managed to get to WP:AIV (I think they first tried facebook, that was rejected, so they left out facebook and tried myspace .. was rejected as well, then moving on for some edits and adding the facebook again .. rejected), the other editor ran into a block for WP:POINT-violations (user logged out to show that IP editors can do good stuff, got blocked after XLinkBot reported them to AIV, unblocked, and then investigated resulting in a block for intentional disruption of the system).

The small group of good faith editors who add a large number of links with which they think they improve Wikipedia generally get the message after being talked to. Those links in the beginning do not end up on XLinkBot, but such editors are being talked to by regulars first. Most of them understand, only every now and then someone panics and needs more significant pushes in the right direction (sometimes they do end up on XLinkBot, especially if they have a wider range of IPS, or if they, in their panic, start socking - it makes clean up and tracking easier, and quick remarks may at a certain point result in them getting the point that we are trying to communicate with them).

But that first group, the true spammers, are, as I said, not here to improve Wikipedia. They are here to improve their own financial situation, or that of a cause they represent, or the financial situation of a company they represent (and there are even 'far fetched' scenarios which nonetheless are happening as well - webmasters who add links to Wikipedia to get more incoming customers to show the resulting web-traffic to their supervisors so they can have a raise, or have better web servers). Do not mistake them for misled regular editors - they are here to make money, they are here to use the free webspace of Wikipedia to improve their situation. They will not stop unless hard forced methods are placed (that starts with XLinkBot, which a lot of spammers simply disregard, but it does have an effect, some do go into discussion and/or stop). The majority of real spammers do everything to stay under the radar. They create numerous socks, they use redirect sites. The only way to stop these are the hard measures of the Spam blacklists and blocking editors. They do not care about your friendly messages (only a strong warning that their site(s) may end up blacklisted may make them consider to back off), they will not be converted. When they can, physically, not spam Wikipedia anymore because their links are blacklisted, the only edits they may do is complain, or try to circumvent the blacklisting. They will not turn into valuable editors - most will simply disappear (but they keep an eye on it .. I have seen cases where spammers got their links blacklisted, and months, years later the links get de-blacklisted and the spammer returns very soon after, it pays to have your links here - the same as schoolkid-vandalism, you see vandalism restart hours or days after the block expires).

Now, the point is going to be, that we need to keep the valuable editors, be friendly enough to them to keep them, and to educate them why certain links are a problem. I strongly believe that XLinkBot has there a task in educating them what may be the problem and asking them to take care next time. Just saying 'I reverted your link, bye' is not enough, there education is a necessity, and also telling them what they can do next. That message may be long (risking tl;dr), but not trying to make them understand the rules is certainly less effective. If they revert the bot not knowing what the problem is, they may (and probably will) be re-reverted by a human editor who will, at the least, ask them why they did not consider the first message, the answer may be 'the first message did not tell me what I did wrong'. It is like a traffic cop who sees a first time offender driving 120 where 80 is allowed, stopping the driver and saying 'I warn you, bye', getting back on his motor bike and driving on, leaving the driver to think 'what did I do wrong' ... In stead, most traffic cops will get off from their motor bike, not even intending to write a fine but with the intention of giving a strong warning, immediately pointing the driver to what they did wrong, and explaining the consequences if it happens again. And I think only very few drivers will get out, throw the keys into the grass and walk home, never to drive again (and they even don't do that when that cop is in a bad mood and does write the fine on the first time).

Currently, our message is aimed at being friendly and giving education, but also at being firm. It is very difficult to guess how many editors walk away after the first remark, but I think that shifting the message (any message) is just going to shift the group. If you give a full explanation, editors will leave because they get reverted and don't want to read it all, if you give a very short message editors will leave because they get reverted and don't get told why. In the end, it is the same percentage that walks away.

Maybe we should have a look at the genuine logged-in new editors, see how many keep editing (IP editors may return as another IP, so it is difficult to get a proper number out of that). Of the editors that are still active (also those behind static IPs), you could ask how they perceived the message that the bot left them - whether they understood it, how they found the tone, etc. etc.

Hope to hear more. --Dirk Beetstra T C 11:59, 24 January 2012 (UTC)[reply]

Thank you for your thoughtful response, Dirk :) I just want to make it clear that when Steven and I approached you (or any other bot op) for a template test, it's not because we thought your templates were bad or wrong. We know you didn't just slap something together in two seconds – you actually put a lot of time and thought into making the message appropriate and helpful (I know this because I had to spend a whole day teasing out all the different parameters and regexes from the config page!). But as you say, it's an incredibly complicated situation: how do you encourage the good editors without overwhelming them with rules and policy, while simultaneously keeping spam out of the encyclopedia? That's just not a problem any one of us could possibly figure out through educated guessing or anecdotal evidence. A/B testing different messages might sound really crude, but over a long enough timeline and with a large enough sample, we actually can say, with statistical confidence and significance, that one template did better than another at retaining users (or getting them to talk to other editors, or getting them to use the help desk...). We have a team of PhD students (User:EpochFail and User:Staeiou) and a data analyst (who helped us raise 20 million dollars for the fundraiser, all through A/B testing!) who can run all the fancy regressions and Chi-squared tests you see in peer-reviewed journals. You can certainly doubt our ability to discern the secret motives that lie deep within the hearts of editors, but I should hope you don't have the same doubts about science and statistics :)
I guess what I'm saying is, we don't have to rely on educated guessing anymore – we have the resources and the staff (thanks in large part to the aforementioned fundraiser, heh) to get scientific evidence about template effectiveness. If the templates Steven and I wrote totally bombed, then it'll be clear that you were on the right track with more explicit explanations for reverts, and we should try retaining that element and testing other variables. Because the other component to all this is that Wikipedia, along with the rest of the Internet, is constantly evolving. What worked six years ago to draw in new people doesn't work now, so the newbies of today and tomorrow will need very different kinds of help and guiding going forward.
Anyway, that's my A/B testing manifesto, hope you enjoyed it :D And thanks again for letting us test with your bot and giving us some meaty food for thought to chew on. We may disagree on a few minor points, but obviously we agree on the basic principles of getting more good editors involved in the project and making Wikipedia even better, and that's all I really care about in the end! Maryana (WMF) (talk) 19:17, 24 January 2012 (UTC)[reply]

Oh, I believe the statistics and the result. The problem is, that I see also the other side of the medal. I do think that our goal should not be only to retain more editors, or only to win more new souls. Sure, every gained editor is one, but that not at the cost of things at the other side. Spammers ánd inappropriate link additions are a real frustration to many, and I'm not sure that we should be happy with retaining one or even 10 editors more, where then more long-term editors start to burn out because of all the work. And that is something that your A/B testing does not measure here. I see the collegue spam-fighters around me, and some take extensive wikibreaks after a fight with the community, or with individuals. And manpower is already thin in this field. The current situation is not good, but changing something here because the statistics say it is an improvement here .. may result in chaos somewhere else. I am very aware of the limits of statistics, it is just what you measure. And if we have a 5% increase in retained editors ánd a 10% increase in re-inserted inappropriate external links, and more downtime of editors cleaning them up because of more work, frustration and burn-outs, and more editors not staying because of all the spam and other rubbish everywhere .. that is not worth it.

I am also not saying that the current message is optimal. I do realize it is long, and people don't read long messages. I would be all for shortening it, or doing something else smart to it. Putting the long message there results in people not reading it (e.g 'I removed the information you added in your edit to the page 'blahdiblah' because in your edit you included an url which points to a document on an site outside of en.wikipedia.org. The movie you are pointing to is an illegal copy of a video clip of the artist blahdiblah, and linking to material which is in violation of the copyright of the original creator of the work is a criminal offense in many countries. Moreover, the url to the movie is inappropriate because you are adding it to page 'blahdiblah', while the videoclip is actually of the clip 'halbidhalb', a performance of 'blahdiblah'. It does not lie in Wikipedia's goal to link to all the videoclips that 'blahdiblah' have produced ...' - tl;dr, after 'I removed the information you added' you stop reading and you re-add the link), putting a short message ('reverted addition of inappropriate external link to copyvio') means that people may not understand why ('reverted', what do you mean? 'inappropriate', who are you to say that? what is an 'external link'? what do you mean with 'copyvio'? - and asking is even more effort than reading a long message, and you have to wait for an answer, so just revert, I mean, copyvio, so what!), and a short message with a link to the full story ('reverted addition of inappropriate external link to copyvio') .. well, editors simply don't follow the link and hence still don't understand, because WP:EL and WP:COPYVIO are too long and it is too difficult to find what part we now actually mean.

Two points of the three are measurable through the A/B test results you have - percentage of retained editors ánd percentage of re-added external links (and you could also measure how many are actually inappropriate external links that are re-added). The third one will be a collateral effect of the latter of the two measured values (with a negligibly small compensation effect of the former) and not measurable through these statistics, but if the percentage of re-added external links is significant, that will also be significant. And over all Wikis (of which en.wikipedia is by far the biggest) we are operating under an addition speed of 1 external link every second - sure, a lot of those are good, but a lot of that still needs to be checked.

Take care measuring retained editors. My experience is that spammers tend to be 'retained editors' until they either get blocked (upon which many sock so they are still a retained editor), or their links get blacklisted. --Dirk Beetstra T C 20:06, 24 January 2012 (UTC)[reply]

About editor retention, see the diffs in time for the page m:User:COIBot/case/case7 .. some editors you don't want to stay. --05:51, 26 January 2012 (UTC)

Linkfarm query re Panthera Hybrids

Hi there, I added a link on the Panthera Hybrids Wikipedia page to an article on the blog of a leading expert on the subject of unusual cat forms after realising from reading this article that the term 'dogla' was being used incorrectly on the Panthera Hybrids Wikipedia page. But you have now removed the link because you say that I had created a Linkfarm. Having read the definition of this, I don't understand why you have done this as his article and blog do not seem to be a Linkfarm. Could you explain your decision so as I can better understand? Thanks very much. 92.13.21.31 (talk) 04:02, 28 January 2012 (UTC)[reply]

Thank you for your question. I will answer this this evening, I am sorry but I am running out of time now. --Dirk Beetstra T C 04:26, 28 January 2012 (UTC)[reply]
Again, thank you for the question. The page you were working on is on Panthera hybrids. A bit of an overview page with a number of pages linked for the specific hybrids. All in all a lot of information. On this page, there are two links with more info, both seem to be authorative sites already. You were now adding a blogspot, and you may very well be right that that contains a lot of information, but I am not convinced at all that the blogspot adds anything substantial that is not already covered by the page, by the other Wikipedia pages already linked on that Wikipedia page, and the two external links. Editors who are reading the Wikipedia page, get already a lot of info, they may have visited a couple of the linked Wikipedia pages on the Wikipedia page, and now may want more info. Then they have to go through 3 external links trying to find info that is not already covered. I would myself first go for the other two sites (and not for the blogspot). In other words, if a link is not adding substantial what is already covered in the Wikipedia, and the other external links, then the link should not be added. Of course, discussion on the talkpage where a careful examination about which site adds more, and which are superfluous would be a good plan. I hope this explains. --Dirk Beetstra T C 17:19, 28 January 2012 (UTC)[reply]

Thanks for your answer, and I certainly take your point about not wanting too many links on the page entry (though it does say at the top of that page that citations are required for it). However, the two links already present are both to the same, fairly old site that perpetuates the misleading definition of the term dogla, whereas the link to the much more recent blog article (written by a recognised, published authority on the subject of unusual cats) that I added specifically exposes this error of definition and explains why it is an error and misleading. That's why I added the link, and why I feel it should stay, otherwise wikipedia readers of the Panthera hybrids page entry will have no idea of the source for this explanation of the correct definition of dogla and also leoger. I shan't try to re-add the link, and in any case i'm new to adding things to wikipedia anyway, but I respectfully suggest that the link should be re-added, for the reasons given here. Thanks.92.13.21.31 (talk) 19:39, 28 January 2012 (UTC)[reply]

You're welcome. First, external links are not the same as references, though they may be suitable to become references later. And for the rest, it is along the lines what I am suggesting: a discussion on the talkpage to see whether the links need updating. Thanks! --Dirk Beetstra T C 20:05, 28 January 2012 (UTC)[reply]

PDB info in Chembox?

Hi Beetstra, The Chembox info is excellent, could PDB info be added to it? ie a link to macromolecular structures which contain a particular molecule. Just a link to the Compound browser would be needed, Choline or biotin to be continuously up to date with the PDB archive. I have a mapping between Inchi and the PDB three letter ligand code which would aid doing it automatically.

Thanks A2-33 A2-33 (talk) 09:22, 28 January 2012 (UTC)[reply]

Would you mind requesting that on the talkpage of the template. I may come to it in the near future, but it may go faster if someone else could do it.
I may however be interested in the file with the links between InChI (StdInChI?) and PDB three letter ligand code (though I can't promise that I have time to use it), could you send me an email via Special:EmailUser/Beetstra? Thanks!! --Dirk Beetstra T C 17:22, 28 January 2012 (UTC)[reply]

Hi, please could you re-check your edit on Aluminium isopropoxide? It is listed on WP:CHECKWIKI, can't solve it by myself - I'm not a chemist. Thanks. --Ben Ben (talk) 14:44, 29 January 2012 (UTC)[reply]

Not sure what happened there, my script must have received broken information from the server it is pulling the data out off. I have retrieved another SMILES code, cleaned up the rest. Thanks for catching this! Happy checking! --Dirk Beetstra T C 14:57, 29 January 2012 (UTC)[reply]

Should CommissionBreakthrough spam be reported?

I manually added an item to WP:WikiProject Spam/CommissionBreakthrough Spam (diff). Do you want people doing that? (If not, please revert.) There's so much of this junk and your system seems to be coping very well, so I'm not sure if new instances should be reported somewhere. In the case that I added, the user does not yet know how to add an external link, so the spam link appears to have never been successfully added. Johnuniq (talk) 06:43, 3 February 2012 (UTC)[reply]

There are some which escape the system, I would say that where you notice that it stays and that XLinkBot does not revert it, please do revert it.
Special:WhatLinksHere/Template:ClickBankSummary is another check, if it is not there (giving COIBot the time to save the report), then it apparently did not get detected. If unsure, just add it I would say. It is really difficult to get a full record of all of this rubbish, people will come back, and people will stay for a long, long time I am afraid. Also, if confident enough with the revertlists, please do add it to the revertlist, the overridelist and the leveloverrule list. Either try to find the unique features, or otherwise just add the domain to all three of them (if you can find the server IP, please add it to the log behind the rule, so I can see that it is there, and that it may need investigation).
Thanks for helping out. --Dirk Beetstra T C 07:23, 3 February 2012 (UTC)[reply]
You can also add detail to the talk page of revertlist and we will get to it. We are needing to watch for zny crosswiki activity so all notes on the talk page are really helpful. — billinghurst sDrewth 09:00, 3 February 2012 (UTC)[reply]

In your notes for blocking this IP/user, you noted the website rankmaniac-2012.caltech.edu. I'm not particularly fond of people abusing Caltech's network; was that the direct site they were spamming, or were they spamming some other site related to the "competition?" If the former, a simple traceroute gives me the owner of the site, and I'm inclined to complain to Dr. Wierman, who is teaching CS 144, which rankmaniac is a part of (sadly, that took far more searching than it should have, and that link is probably one of the only legitimate links using that term anywhere...).--Constantine (talk) 01:52, 4 February 2012 (UTC)[reply]

Hi, thanks for that. The situation is more complex than that. I have blacklisted a number of sites on meta:
  • \b4\.bp\.blogspot\.com\/-Lujq9b278Lg\/TyeYmk66isI\/AAAAAAAAAAc\/yELwrwzyVeQ\/s1600\/caltech_rankmaniac_2012\.gif\b
  • \brankmaniac2012\.webs\.com\b
  • \brankmaniac2012caltech\.blogspot\.com\b
  • \bmyrankmaniac2012\.blogspot\.com\b
  • \brankmaniac2012caltech\.tumblr\.com\b
  • \bcs144rankmaniac2012\.blogspot.\com\b
And that shows a big part of the problem. Websites are easy to get, if that was different, this would have solved it, but it is a matter of either being creative (you can see already three blogspots ..), or finding one of the thousands of free webhosts out there. They did not abuse the caltech.edu site itself (that came out of a websearch for the class, I thought it was the official competition site), probably that is not the site they should promote (but obviously they would abuse the CalTech network, but that was their assignment). Their SEO does their job then, a lot of illegitimate use of the term.
As you saw, I blocked two accounts indef, and 4 CalTech IPs for a month. The IP blocks may upset other editors, but generally IPs in educational institutions are static, it is likely that if they used a private computer for the spamming I actually hit the right editors there, or otherwise I hit a number of computers used in the class.
Others have salted the page on Wikipedia, which I think is a good plan as well to keep the spamming away.
I'll not explain my reasoning further (per WP:BEANS), but I do think (and I have been working for years fighting spam on Wikipedia), that they all did fail their class massively - they got their work blacklisted. If you have further questions, don't hesitate to ask. --Dirk Beetstra T C 04:16, 4 February 2012 (UTC)[reply]
Ah, I see you are using the CalTech network as well. It is good to see that now we know that one of the logged in users used that IP that I block directly. Do you still have problems with the autoblock, it seems you changed IP.
By the way, I brought this to AN: Wikipedia:Administrators' noticeboard#CalTech: Rankmaniac 2012. --Dirk Beetstra T C 04:24, 4 February 2012 (UTC)[reply]
Add rankmaniac2012.webs.com to the list. --Dirk Beetstra T C 05:47, 4 February 2012 (UTC)[reply]

Disambiguation link notification

Hi. When you recently edited Get Rich Click, you added a link pointing to the disambiguation page The View (check to confirm | fix with Dab solver). Such links are almost always unintended, since a disambiguation page is merely a list of "Did you mean..." article titles. Read the FAQ • Join us at the DPL WikiProject.

It's OK to remove this message. Also, to stop receiving these messages, follow these opt-out instructions. Thanks, DPL bot (talk) 10:38, 5 February 2012 (UTC)[reply]

FYI - I just cleaned up the article on Thomas Milton Tinney, Sr. per the tags you previously added. (See also [2].) I'd appreciate it if you'd have a look. After deleting all the cruft, it appears to me that this fellow is downright unnotable, despite the fact that he's listed in several Who's Whos (perhaps they were the pay-to-play variety), so I added a BLP deletion tag. I'd appreciate it if you'd follow the article as I don't edit here very often. Thanks! 75.6.11.19 (talk) 16:07, 7 February 2012 (UTC)[reply]

Thank you. I am not a specialist in the article, I think someone more knowledgeable should look at it. I think that some of the problems are still there, so the tags should probably still stay. Thanks! --Dirk Beetstra T C 16:10, 7 February 2012 (UTC)[reply]
I can tell you for a fact that Tinney is non-notable in his field (genealogy), but of course, that's just original research. A Google search produces absolutely nothing written about him, just promotional material and message board posts. But I don't think his biography is particularly "specialist" material. It was written in a pretty quirky way, which made it hard to discern exactly what was meant, so I may have thrown some of the baby out with the bathwater. So I'd be happy if anyone other than Tinney had a look at it. If no one does, then I guess he's non-notable. 75.6.11.19 (talk) 16:38, 7 February 2012 (UTC)[reply]

MSU Interview

Dear Beetstra,


My name is Jonathan Obar user:Jaobar, I'm a professor in the College of Communication Arts and Sciences at Michigan State University and a Teaching Fellow with the Wikimedia Foundation's Education Program. This semester I've been running a little experiment at MSU, a class where we teach students about becoming Wikipedia administrators. Not a lot is known about your community, and our students (who are fascinated by wiki-culture by the way!) want to learn how you do what you do, and why you do it. A while back I proposed this idea (the class) to the community HERE, were it was met mainly with positive feedback. Anyhow, I'd like my students to speak with a few administrators to get a sense of admin experiences, training, motivations, likes, dislikes, etc. We were wondering if you'd be interested in speaking with one of our students.


So a few things about the interviews:

  • Interviews will last between 15 and 30 minutes.
  • Interviews can be conducted over skype (preferred), IRC or email. (You choose the form of communication based upon your comfort level, time, etc.)
  • All interviews will be completely anonymous, meaning that you (real name and/or pseudonym) will never be identified in any of our materials, unless you give the interviewer permission to do so.
  • All interviews will be completely voluntary. You are under no obligation to say yes to an interview, and can say no and stop or leave the interview at any time.
  • The entire interview process is being overseen by MSU's institutional review board (ethics review). This means that all questions have been approved by the university and all students have been trained how to conduct interviews ethically and properly.


Bottom line is that we really need your help, and would really appreciate the opportunity to speak with you. If interested, please send me an email at obar@msu.edu (to maintain anonymity) and I will add your name to my offline contact list. If you feel comfortable doing so, you can post your name HERE instead.

If you have questions or concerns at any time, feel free to email me at obar@msu.edu. I will be more than happy to speak with you.

Thanks in advance for your help. We have a lot to learn from you.

Sincerely,

Jonathan Obar --Jaobar (talk) 15:47, 8 February 2012 (UTC)[reply]

Reverting Edit Versions on article "St. Peter's College, Colombo 04"

Dear Mr.Dirk Beetstra,
I am a web team member of St. Peter's college, Colombo 04. As per orders from the administration, i have been requested to continuously update this article from this year onwards.

Being a Web Team member, it is my duty to update information on various clubs and societies, and also all the activities carried out by those respective bodies. Since there have been many changes in their organizational structures and since some clubs have combined together, i have had to edit the article respectively.

Please be kind enough to allow my editing as it is being done with due permission.

You may reach me at <redacted, email removed>— Preceding unsigned comment added by teen95 (talkcontribs)

Thank you for the remark. Unfortunately, that does not mean that you should not follow our policies and guidelines, and actually, you should be editing more careful than a regular editor, because you have a conflict of interest. --Dirk Beetstra T C 05:00, 9 February 2012 (UTC)[reply]

UTRS Account Request

I confirm that I have requested an account on the UTRS tool. Dirk Beetstra T C

Hi Beetstra, thank you for your interest in the tool. I've approved your account, please feel free to login and test the system.
As part of this beta test, we'd like everyone to test every aspect of the tool. This includes acting as blocked users - we'd like each of you to file at least two appeals and respond to them as though you are blocked. Please try to act like a blocked user new to Wikipedia, unfamiliar with common terms and probably a bit frustrated at the situation.
When reviewing appeals, please act as though you are reviewing real blocks. You should be able to comment on any appeal, regardless of who has reserved it; reservations only ensure that reviewers don't send conflicting emails.
If you encounter any bugs (things not appearing to work right, and especially error messages), please file a bug report on JIRA. You will need to register an account there. New features can be suggested there as well, but please add the "after-beta" label to these so we can easily prioritize between bugs that must be fixed and features that can be added later.
Thank you again for volunteering to beta-test.--v/r - TP 00:35, 12 February 2012 (UTC)[reply]

Filton College

I am in the process of rewriting the article so please can you let my finish and then change it. Mark999 (talk) 11:26, 9 February 2012 (UTC)[reply]

Could you please not post telephone numbers, and inappropriate external links. We are not a linkfarm or the yellow pages. In that case I do not need to clean up the article after you. Thanks. --Dirk Beetstra T C 11:27, 9 February 2012 (UTC)[reply]

360cities.net

360cities.net: Linksearch en (insource) - meta - de - fr - simple - wikt:en - wikt:frSpamcheckMER-C X-wikigs • Reports: Links on en - COIBot - COIBot-Local • Discussions: tracked - advanced - RSN • COIBot-Link, Local, & XWiki Reports - Wikipedia: en - fr - de • Google: searchmeta • Domain: domaintoolsAboutUs.com

Hi Dirk,

Am currently aiming at extending the Latvian (mainly nature) articles in French wiki, just started on Parc national de la Gauga in French wiki. Found a beautiful 360° panorama of it in Gauja National Park en.Wiki.org ( titled 'Tourist trail in Gauja National Park' in section 'External links'). Can't put the 360cities link here, it's blocked). Looks like the block was addes by COIBot, yr bb. Had a quick check on the net, don't see any major pbms wth that site. I opened it a few days ago, have had no infection, spam nor any such thing despite my computer using only basic protections. That site looks safe enough to me. Why is it blocked in frnch, what's wrong with that site, & how to unblock it if poss ?

Another weird thing is, your talk page somewhere else (there : http://en.wikipedia.7val.com/w/index.php?title=User_talk:Beetstra&action=edit&section=new) says my IP is blocked - apparently by some 'zzuuzz' dude (??? Page mentioning the block). Signalling it despite being able to reach you here, as I don't know if there isn't any link between the 2 events. Thanks for answer. Fran90.8.241.153 (talk) 12:08, 12 February 2012 (UTC)[reply]

Regarding the first, I see it is blacklisted on several wikis, it, es and zh wikipedia, and it is listed on meta. That generally suggests that it has been abused by editors (spammed) - especially since it was first tried to be stopped on three individual wikis before global blacklisting was implemented. Note that that does not mean that the site itself is containing anything bad. COIBot only reports it, it are human editors who check whether things should be blacklisted. I guess you would have to ask on meta (see  Defer to Global blacklist - I don't think it stands much of a chance), or ask whitelisting for the specific link on the wiki where you want to use it (here on en:  Defer to Whitelist, but you'd have to be on the fr wiki).
The wiki you are referring to does not seem to be a mediawiki wiki (it reads '7val.com' in the url). I think that I would not get to see if you leave me a message there. The site is strange, it has a strange display.
I hope this helps. --Dirk Beetstra T C 12:40, 12 February 2012 (UTC)[reply]
I believe that 7val is a self-appointed mirror of Wikimedia and a bit more, how much more, is unknown. Look at en.wikipedia.7val.com, en.wikisource.7val.com and others. Not sure why we would host links to it of our own data. — billinghurst sDrewth 13:06, 12 February 2012 (UTC)[reply]

Twitter references and ITV

The problem is not that the information is not there - the problem is you don't understand the information that is provided. Glos in an abbreviation for Gloucestershire and that is known throughout the UK, so why would someone write Gloucestershire when they are limited on how many characters they can type?

ITV regional reporters give their Twitter IDs every time they appear on TV and ITV use the Twitter feeds as a cheap way of providing updates due to regular technical problems with their regional websites. It is easy to identify the official feeds with reliable information this way. The only alternative to Twitter for sourcing such information is LinkedIn.

89.248.29.41 (talk) 13:06, 14 February 2012 (UTC)[reply]

Yes, I see the whole feed, and you say that one person is pregnant. I did not see that post. First, it is a primary source, if and only if that is really the official twitter, which I will assume, secondly, it is simply not a reliable source. Please read the guidelines. --Dirk Beetstra T C 13:09, 14 February 2012 (UTC)[reply]
No - one person is not pregnant, the person has recently given birth and is on maternity leave to be able to spend time with their new born. That's exactly what I mean about you not seeming to understand what's been written.
Wikipedia guidelines state the following:
Self-published and questionable sources may be used as sources of information about themselves, usually in articles about themselves or their activities, without the requirement in the case of self-published sources that they be published experts in the field, so long as:
1. the material is not unduly self-serving;
2. it does not involve claims about third parties;
3. it does not involve claims about events not directly related to the source;
4. there is no reasonable doubt as to its authenticity;
5. the article is not based primarily on such sources.
This policy also applies to pages on social networking sites such as Twitter, Tumblr, and Facebook.
Therefore, I don't see why Twitter references can't be used in such circumstances. I suggest you familiarise yourself with Wikipedia guidelines before you starting removing information, saying it doesn't comply.

89.248.29.41 (talk) 14:03, 14 February 2012 (UTC)[reply]

Doesn't matter, the feed still did not show that, and still it is better to have an independent source for it. --Dirk Beetstra T C 14:10, 14 February 2012 (UTC)[reply]
The writer of that twitter feed (and I'll assume it is the person who you are talking about in the Wikipedia text), does not say anything about being on being on maternity leave .. the closest it gets is that she twitters that she misses the company she works for .. --Dirk Beetstra T C 14:18, 14 February 2012 (UTC)[reply]
The fact that she's on maternity leave was already on the Granada Reports page before I edited it and wasn't referenced - you will see this if you look at the edit history. Her return date wasn't mentioned on the page - I added this and referenced her Twitter page as the source FOR THE RETURN DATE - NOT FOR THE CLAIM OF HER BEING ON MATERNITY LEAVE.
The Twitter feed says "@lucymitv @GranadaReports miss you too! Should be back in April!" in response to Lucy Meacock's tweet of "@KeriEldridge great to hear from you. When are u back on our screens? @GranadaReports missing you!" - Granada Reports being the name of the program the article is about!
If you go to the Granada Reports Facebook page: http://www.facebook.com/granadareports and look at all the posts for the past 10 months you'll see there was one from Keri Eldridge after she did her last breakfast bulletin before going on maternity leave. Of course, if people like you didn't remove references to social networking sites the two references together would be together and provide a reference for both her being on maternity leave and her expected return date.
Quite frankly you've been quite pedantic and it seems you've just tried to waste my time to be awkward and get attention. As I've already told you THERE IS NO INDEPENDENT SOURCE AVAILABLE. I think you should avoid editing pages related to ITV regional news programmes, you are only going to annoy the people who work hard to ensure correct information is available from very limited resources with your approach and that will only result in the content available on Wikipedia being less useful. 89.248.29.41 (talk) 16:28, 14 February 2012 (UTC)[reply]
'Should' .. wow, that is making it really reliable. And are you sure that is the return date for her maternity leave? The post does not say it.
So, there is no independent source available. It must be a very interesting, crucial piece of information then. We are writing an encyclopedia here. --Dirk Beetstra T C 16:31, 14 February 2012 (UTC)[reply]

On that basis 99% of the content of the 'Granada Reports' article needs to be removed immediately because only place most of the content can be verified is via a link updated by ITV or a journalist employed by ITV. I trust you will be doing that immediately, since you care so much about the issue. Why can't you follow Wikipedia guidelines and add [citation needed] tags where something needs a better reference? 89.248.29.41 (talk) 17:30, 14 February 2012 (UTC)[reply]

I am following our policies - this info is not notable, it should have been removed altogether. That other stuff is there is not a reason to keep this or add more, see WP:OTHERCRAPEXISTS. Thanks for notifying me though that you brought the issue to AN/I. --Dirk Beetstra T C 17:37, 14 February 2012 (UTC)[reply]

Cryptic?

Not sure what you meant by the "Wow, that is quite cryptic again" jab on ANI, but if I am not being clear, please let me know. Don't give snarky comments, let me know what I am not being clear about. - NeutralhomerTalk18:49, 14 February 2012 (UTC)[reply]

Oh, Neutralhomer, it was not about you. I see now that my post is misformatted. Sorry for the misunderstanding. I was talking about the post on Twitter, which was mentioning 'glos' for 'Gloucestershire'. No, you were very clear, I just did not understand that I missed that the twitter link did mention Gloucestershire, since I did search/check for it. --Dirk Beetstra T C 19:03, 14 February 2012 (UTC)[reply]

Far-left politics

Hello Beetstra,

I have seen that you have reverted on the article Far-left politics without giving an explanation in the edit summary or participating in the discussion on the talk page. I am sure that you understand that this topic is quite touchy and I want to inform you that there has an edit war been going on. I don't think that reverting without contributing to the discussion is the right thing in this situation. It could fuel the edit war and does absolutely not lead to a solution of the disagreement. --RJFF (talk) 19:33, 15 February 2012 (UTC)[reply]

Yes, I saw the edit war, and I blocked the editor edit warring. I am intentionally NOT participating in the discussion, I do however see that editors are reverting over and over - if I see further removals without that the talkpage discussion has come to a conclusion, I will take more drastic steps. Thanks. --Dirk Beetstra T C 20:18, 15 February 2012 (UTC)[reply]
Do note that I blocked Spylab for 1 week for (slow) edit warring. I hope that I do not need to hand out more blocks. --Dirk Beetstra T C 20:20, 15 February 2012 (UTC)[reply]

Disambiguation link notification

Hi. When you recently edited Victor van Amerongen, you added a link pointing to the disambiguation page Space City (check to confirm | fix with Dab solver). Such links are almost always unintended, since a disambiguation page is merely a list of "Did you mean..." article titles. Read the FAQ • Join us at the DPL WikiProject.

It's OK to remove this message. Also, to stop receiving these messages, follow these opt-out instructions. Thanks, DPL bot (talk) 10:28, 17 February 2012 (UTC)[reply]

Ergoloid

Hi Dirk! Could you review this edit of yours to Ergoloid? Your additions seem to refer to dihydroergotamine mesilate, while ergoloid mesylates are a combination with three other (closely related) substances. Cheers, ἀνυπόδητος (talk) 09:52, 18 February 2012 (UTC)[reply]

Another one: [3]. I don't think you should take InChIs from PubChem without reviewing, especially for polymers, as PubChem frequently lists the polymer's components instead of the whole thing. See the structure drawing in [4]. --ἀνυπόδητος (talk) 09:59, 18 February 2012 (UTC)[reply]

Okay, I'll be silent in a moment, but what is this? A bug in your script? --ἀνυπόδητος (talk) 10:01, 18 February 2012 (UTC)[reply]


Regarding the first one, that must have gotten IDs from somewhere, while it absolutely should not get a CSID and other IDs. Those fields should be blank. If the fields are there, I extrapolate the rest of the data from there.
The second one is a valid one, there is nothing wrong with those InChI's.
The third one is a bug in the page, which propagates.
No, please, go on, I'm happy someone is reviewing my edits, but do realise that sometimes there are already errors in the pages, which simply propagate. --Dirk Beetstra T C 10:07, 18 February 2012 (UTC)[reply]
I have updated the first one, also off-wiki for my scripts, it should now update to a CSID = NA - this compound should just not display a CSID.
I have also solved the 'bug' in the last one, I've removed all the duplication there. --Dirk Beetstra T C 10:18, 18 February 2012 (UTC)[reply]
I re-added the CAS no. of ergoloid as [5] explicitly states this number refers to the mixture. Please correct me if I'm mistaken.
Regarding colestipol: I'm not good at reading InChIs, but that one seems to contain a chlorine which colestipol doesn't. See the structure in the article. --ἀνυπόδητος (talk) 10:25, 18 February 2012 (UTC)[reply]
Hmm, ChemSpider for Colestipol really seems about that compound, and gives those InChI's. I think that InChI sees polymers as some form of 'mixture' of monomers? I'll add Colestipol to an ignore list. We can always verify that one manually.
Oops, the CAS is an oversight, indeed, that one can be correct, they do mixtures in a better way.
Again, please keep reporting if and when you find 'errors', in the end we will only get better because of this. --Dirk Beetstra T C 10:32, 18 February 2012 (UTC)[reply]
Re: InChI's and polymers - It is not possible to generate an InChI for a polymer (at this time). Some databases contain InChIs for the monomers of the polymer (or other similar fragments) - I'm not certain of the reasons for this - It could be an intentional approach to try and capture some structural information - in the absence of a polymer InChI, or a case that the database uses name-to-structure algorithms across their data to generate structures (and afterwards InChIs) and that the name has given rise to these molecular fragments. --The chemistds (talk) 22:15, 19 February 2012 (UTC)[reply]
ChemSpider (and PubChem) also give the SMILES, the chemical formula, the structure, and the molecular mass for the monomers. It's definitely incorrect that colestipol has a mass of 281. I fear that ChemSpider and PubChem simply don't handle polymers correctly. --ἀνυπόδητος (talk) 11:00, 18 February 2012 (UTC)[reply]
Strangely, most polymers do not have a record on chemspider (polyethylene, polypropylene, etc.). I am actually surprised to see this one, maybe it is a rare exception. I see as an only solution to do those manually, if my script is doing its job correctly, it should be ignored now. Otherwise I will bash it in place. --Dirk Beetstra T C 11:02, 18 February 2012 (UTC)[reply]
ChemSpider is a database of small molecules - currently the database does not support storing polymers or other extended materials. The database is generated by aggregating data from over 400 datasources. All data is imported based on structures and then any associated data (which may include names). Therefore, it is possible for a datasource to supply a structure with an incorrectly associated name. Looking at the records that have been highlighted - I would summise that that the data probably came from PubChem originally. --The chemistds (talk) 19:56, 19 February 2012 (UTC)[reply]
Other examples are colestilan which resolves to "2-(Chloromethyl)oxirane - 2-methyl-1H-imidazole (1:1)" and colesevelam which resolves to "6-(allylamino)hexyl-trimethyl-ammonium; N-allyldecan-1-amine; 2-(chloromethyl)oxirane; prop-2-en-1-amine; chloride; hydrochloride"; possibly they only list polymers that are drugs. I've just removed SMILES, InChIs and formula from colesevelam. --ἀνυπόδητος (talk) 11:17, 18 February 2012 (UTC)[reply]

Hmm .. this is annoying, and I am really sorry for causing this trouble. I encountered a good handful of cases which simply do not have a proper InChI etc., like the polyolefins etc., but which were nicely out of ChemSpider and hence did not make problems. Now we have suddenly a group where ChemSpider is problematic. I did consider that we might have an odd case where we do have typos in the system and hence get a mistake, but this is heavily thwarting our work. I have stopped working with the script for now - either we need to figure out an easy workaround, or we for now we have to accept a bit of 'errors', and we solve those as we go by. But first, can you think of polymers that should be excluded here, is it easy to make a cross-section of drugs with 'polymer', so we can just make the script ignore these?? --Dirk Beetstra T C 11:29, 18 February 2012 (UTC)[reply]

Sorry, can't think of anything at the moment. And neither dextran [6] nor starch [7] are strictly drugs. (See the Names and identifiers sections in the ChemSpider links.) --ἀνυπόδητος (talk) 14:23, 18 February 2012 (UTC)[reply]
As both a member of staff working on ChemSpider - and someone curates in their free time I can say that ChemSpider (and I) take data quality very seriously. Part of the way that we do this is by employing data validation processes. One key aspect is getting users to approve names that they know are correctly associated with a structure, if a name is highlighted in bold face it has been verified and should be correct, if a name is only formatted in normal face it may be correct but there is a possibility that it has been associated incorrectly. In the cases that are identified above the names were included in the record but were not verified.
With regards to resolving the issues with external data - in the case of ChemSpider I would encourage you to feedback when you discover data that you believe is incorrect. This is central to the ethos of ChemSpider - we know that there can be issues with the data but think that users don't have to just accept it - you can make a difference and get the issues fixed. This can be a simple as using the Leave Feedback button on the ChemSpider record - Or if you create an account, you can add data or mark it for deletion (giving you the ability to interact with the record much like Wikipedia enables you to do). If you wish to know more please get in touch with me or the rest of the team and we'll be happy to discuss this further. --The chemistds (talk) 21:46, 19 February 2012 (UTC)[reply]

As long as they do not have a CSID, it is fine (we can always set it to 'NA' to 'block' it). I may start running it normally again tomorrow. I have a long list, may try to do an hour or two of CSID-validation this evening. Thanks for the reports and the help! --Dirk Beetstra T C 14:26, 18 February 2012 (UTC)[reply]

Is this edit on Triptorelin caused by the same problem? The InChI is definitely wrong. Daniel Bonniot de Ruisselet (talk) 15:40, 18 February 2012 (UTC)[reply]

Wrong input for the script, most of the identifiers that were there were totally wrong. I have removed all of them, will also go through my input files to see if I have it in there somewhere. Thanks for catching this one. --Dirk Beetstra T C 16:31, 18 February 2012 (UTC)[reply]
Thanks. Only this edit added back the wrong data. Daniel Bonniot de Ruisselet (talk) 12:47, 20 February 2012 (UTC)[reply]
I can't see any point where the name Triptorelin has ever been associated with 1,4-piperazine - so I think this might be data from some other source. --The chemistds (talk) 13:27, 20 February 2012 (UTC)[reply]
I don't know where this list is coming from, but it has the following:
  • 13835351Toluene_diisocyanate
  • 13835401Diethylenetriamine
  • 13835459Triptorelin
  • 13835550Ethylenediamine
  • 13835557Pyrogallol
(>5700 records in total). Either I made a mistake, or I got this list with an error. Does not matter, it needs to be resolved, I have removed the record, and double-checked whether it is nowhere in my lists. I hope it is resolved now. --Dirk Beetstra T C 13:52, 20 February 2012 (UTC)[reply]
It does not seem to be 17290424 either, is it? --Dirk Beetstra T C 13:56, 20 February 2012 (UTC)[reply]
I'm not certain at this time (CSID 17290424 is a strong candidate for the correct record)- I need to find some reliable sources that depict the structure fully - or alternatively construct the peptide myself from the appropriate amino acids. I hope to get the chance to do this in the next day or so. The big issue is ensuring that not only is the general structure correct but that all stereocentres are the correct configuration. I'd not the that current wikipedia image contains one undefined stereocentre - which I'm fairly certain is incorrect. You also need to bear in mind that Triptorelin seems to be interchangeably referred to as Decapeptyl and Gonapeptyl which are brand names that I believe relate to formulations containing specific salts (see the talk page for Triptorelin. --The chemistds (talk) 14:16, 20 February 2012 (UTC)[reply]
For me the giveaway was the first stereocenter mentioned in the name, Wikipedia seems to have a D-enantiomer, ChemSpider talks of L- (5-oxo-D-prolyl-L-histidyl-Ltryptophyl-L-seryl-Ltyrosyl-3-(1H-indol-2-yl)-L-alanylleucyl-L-arginyl-L-prolylglycinamide vs. 5-Oxo-L-prolyl-L-histidyl-L-tryptophyl-L-seryl-L-tyrosyl-D-tryptophyl-L-leucyl-L-arginyl-L-prolylglycinamide) .. --Dirk Beetstra T C 14:19, 20 February 2012 (UTC)[reply]
Possibly, but names are only as good as the person (or software) that generated them - (and the source that they used for the structure). I'd also ideally like to see some primary literature to confirm the structure. I'd note that the Merck Index 14th Ed does not indicate that the (oxo-)Proline is the D enantiomer - by convention this would imply that it is the L-enantiomer. --The chemistds (talk) 14:35, 20 February 2012 (UTC)[reply]
We're for sure not at 100% overlap yet. Maybe this is a missing record, maybe there is something wrong on either end. I do pick out sometimes cases where there are real mismatches, even while maybe the link is correct. I find sometimes records where some things don't match up, but where I don't know where. I just tend to blank the CSID on-wiki, we'll get to it later.
Bit of a side, do you by any chance have access from ChemSpider side to a list of ChemSpiderIDs with confirmed Wikipedia pagenames, or could you generate that. I once got one such list, but I guess both of us have significantly progressed.
Our coverage of (hopefully correct) CSID's is at the moment at 4627 out of 5122 pages with a drugbox (90.33%) and 7116 out of 7780 pages with a chembox (91.46%), so that is 11743 out of 12902 pages - about 1200 missing links. Here is a long, long list of almost 1000 pages with their current CSID, but for which I have no clue whether they are correct - they need to be checked by eye. I hope I will have some time to go through it, and make a new list 'CSID=Pagename' (which I can then feed into my script). The remaining 300 probably currently do not have a CSID at all (or are cases where there is no match, perhaps). --Dirk Beetstra T C 16:26, 20 February 2012 (UTC)[reply]

I am going to restart the script again, please do poke me when there are articles mishandled, whether it is from wrong input from the page (wrong identifiers already there), whether it is from wrong data externally (the external database says that compound a has identifier 1, but that is wrong), massive blunders from my side (I do make them), improper data externally (like the InChI's for the polymers), or a combination of some or all of these. --Dirk Beetstra T C 08:01, 19 February 2012 (UTC)[reply]

Hmm, quite a lot of different things to address here. I'll add comments in relevant places above. --The chemistds (talk) 19:24, 19 February 2012 (UTC)[reply]

ChemSpider links

Hi Dirk, Following up on the discussion above. We should be able to supply an updated list of ChemSpider records with Wikipedia links. I think that it would be useful if you can supply a list of Chembox/drugboxes that have a CSID, InChI or SMILES - it would help us to identify where we are not currently providing links in to Wikipedia.

Looking at the list in your sandbox - I'd point out that there is a block of ~30 entries that are actually User pages.

With respect to the figures that you give for the number of drug/chem boxes that are missing links ~1200 - It would be nice if this is the case, but I'm not sure how to rationalise this with the number of pages that are missing a CSID - currently listed as ~3700. Some of this may be due to the listing of User pages on the Category:Chemical_pages_needing_a_ChemSpiderID, another part of the disparity may be that some ChemBoxes have several CSIDs - for different enantiomers, my guess is that there are still somewhere between 2500 and 3000 Chem/drugboxes that are missing a link to ChemSpider - but I'll be happy to be proved wrong. --The chemistds (talk) 21:33, 20 February 2012 (UTC)[reply]

No, I'm wrong, I should go to be earlier. I forgot that when a chembox/drugbox is void of certain data it is also 'verified' - if a page has a correct KEGG and the rest of the identifiers blank and recording that revid will result in CheMoBot only tagging the KEGG. The other fields are wrong, but since they are empty, CheMoBot will not tag them (or the tag will be invisible). When someone adds a, say, ChemSpiderID to it, the CSID may be correct, but since it is not verified to be correct, CheMoBot will tag it with a ☒N until someone verifies if that is the correct ChemSpiderID.
Regarding the userpages, we can just ignore them, update them, or index them, whatever - of course taking into account what a the 'owner' wants with it.
I'll see if I can convince the script to spew out a list of wikipedia data. --Dirk Beetstra T C 04:11, 21 February 2012 (UTC)[reply]

Dan Ariely

Hi Beetsra,

I added a link to Dan Ariely's external links, but I noticed you removed it - can you pls tell me why?

I am relatively new to editing and the laws of wikipedia, so I am keen to find out.

Thanks — Preceding unsigned comment added by Jamesstatham (talkcontribs) 17:26, 21 February 2012 (UTC)[reply]

Thank you for your question. I'll try and explain. First of all, we are writing an encyclopedia here, not a linkfarm or an internet directory. Pages should be as much as possible self-contained. However, Wikipedia can not incorporate all information. Therefore, we can incorporate external links. However, it lies not in Wikipedia's goal, to link to every resource about a subject. If I see the page you were linking to, I do not see a lot of information that is not already in the Wikipedia page, moreover, most of it could be incorporated easily. Then, you use as a description of the link "Represented for speaking engagements by Leigh Bureau" - that looks certainly promotional. Were you linking to the site to give more information about the subject (which I don't think the site has), or were you linking to promote (I do not regarding the latter, that you first change an existing link, then change the text, and then move the link higher up so it looks more prominent. I would really suggest that you heed the warning that Dmacks left you. --Dirk Beetstra T C 18:31, 21 February 2012 (UTC)[reply]