Jump to content

Wikipedia:Village pump (policy): Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Line 1,465: Line 1,465:
:This happens due to a bad rule that we have, called [[WP:PLOTSOURCE]]: {{tqq|The plot summary for a work, on a page about that work, does not need to be sourced with in-line citations, as it is generally assumed that the work itself is the primary source for the plot summary.}} This is one of the few areas of Wikipedia where--for reasons I [[WP:WALLEDGARDEN|can guess]]--we have a guideline that violates a policy ([[WP:NOR]]). Works of fiction are primary sources for their own plots, and normally we don't allow editors to interpret primary sources because that's original research. Notable works of fiction has [[WP:RS]]es about them, of course (reviews), and those include plot summaries. What we ''should'' do is require all plot summaries in Wikipedia articles to be summaries of plot summaries written by RSes. That way, we can be sure that it is the RSes, and not editors, who are interpreting the plot. "Interpreting the plot" means summarizing it, identifying important plot points, conveying how they are related, etc., all of which should be done by RSes. It's not a surprise, then, that when editors summarize a plot by themselves (instead of summarizing RS plot summaries), they end up summarizing the plot wrong. [[User:Levivich|Levivich]] ([[User talk:Levivich|talk]]) 16:22, 6 February 2023 (UTC)
:This happens due to a bad rule that we have, called [[WP:PLOTSOURCE]]: {{tqq|The plot summary for a work, on a page about that work, does not need to be sourced with in-line citations, as it is generally assumed that the work itself is the primary source for the plot summary.}} This is one of the few areas of Wikipedia where--for reasons I [[WP:WALLEDGARDEN|can guess]]--we have a guideline that violates a policy ([[WP:NOR]]). Works of fiction are primary sources for their own plots, and normally we don't allow editors to interpret primary sources because that's original research. Notable works of fiction has [[WP:RS]]es about them, of course (reviews), and those include plot summaries. What we ''should'' do is require all plot summaries in Wikipedia articles to be summaries of plot summaries written by RSes. That way, we can be sure that it is the RSes, and not editors, who are interpreting the plot. "Interpreting the plot" means summarizing it, identifying important plot points, conveying how they are related, etc., all of which should be done by RSes. It's not a surprise, then, that when editors summarize a plot by themselves (instead of summarizing RS plot summaries), they end up summarizing the plot wrong. [[User:Levivich|Levivich]] ([[User talk:Levivich|talk]]) 16:22, 6 February 2023 (UTC)
::I agree, but the headwind on this is very strong. [[User talk:Donald Albury|Donald Albury]] 16:28, 6 February 2023 (UTC)
::I agree, but the headwind on this is very strong. [[User talk:Donald Albury|Donald Albury]] 16:28, 6 February 2023 (UTC)

== RFC: Clarifications to [[WP:COSMETICBOT]] for fixing deprecated HTML tags ==

This is a request for comment regarding [[Wikipedia:Bot policy]], specifically the section on restricted bot tasks related to cosmetic changes outlined at [[WP:COSMETICBOT]]. Ultimately, I would like to start a discussion about whether or not the bot account [[User:MalnadachBot]] is operating within Wikipedia policy, and whether [[WP:COSMETICBOT]] needs to be updated to clarify how these kinds of bots should be handled in the future. First, a bit of history behind this situation:

{{User|MalnadachBot}} is a bot account that is owned by [[User:ಮಲ್ನಾಡಾಚ್ ಕೊಂಕ್ಣೊ]]. This is an [[WP:AWB|AWB]] bot that primarily works on fixing [[Lint (software)|lint errors]], which are common technical issues in wikitext markup. A list of lint errors detected on Wikipedia is available at [[Special:LintErrors]]. The bot has had a total of 13 tasks approved by the Bot Approvals Group. The first 11 tasks were fairly standard error-fixing tasks, each of which had a limited scope to fixing a specific type of error. [[User:MalnadachBot/Task 12|Task 12]] was requested to fix another specific error associated with deprecated <nowiki><tt>...</tt></nowiki> HTML tags. However, the BAG member who approved the task (in this case, [[User:Primefac]]) decided to expand the scope of the task to fixing ''all'' linter errors across Wikipedia, and speedily approved the task. Since that task approval, MalnadachBot has made millions of edits, becoming one of the top accounts by edit count in Wikipedia history.

One of the most common edits that the bot makes is to replace {{xt|<nowiki><font color="red">...</font></nowiki>}} tags with {{xt|<nowiki><span style="color:red;">...</span></nowiki>}} tags. This is because <nowiki><font></nowiki> tags were technically deprecated when HTML5 was introduced, so current best practices for writing new HTML are to use span tags instead of font tags for customizing the font and style of text. If you inspect the bot's contributions, you'll notice that the vast majority of its edits are in the Wikipedia, Wikipedia talk, User talk, and Article talk namespaces. This is because using custom font styles is relatively uncommon in WP articles themselves. However, they are very common in many user's signatures (including my own). So, MalnadachBot is essentially going through nearly every discussion that has ever taken place in the 20+ year history of Wikipedia, and updating old user signatures to use proper HTML. Many of these discussion pages have not been edited in 10+ years, and many of them receive nearly zero annual pageviews. It has made millions of such edits, and there are many millions more in the queue to be done. Over the past 12 months that this task has been running, many users have complained on the bot owner's talk page and [[Wikipedia:Bots/Noticeboard]] that the bot is clogging up watchlists and revision histories with millions of trivial, purely cosmetic edits, but there has been no wider community discussion to come to a consensus on how to move forward. If you used to use <nowiki><font></nowiki> tags in your signature, you've no doubt had your watchlist bombed by MalnadachBot in the last year.

One thing to note is that these kinds of edits do not result in a change to the way that these pages are rendered to the user. This is because, despite <nowiki><font></nowiki> tags being deprecated in HTML5, ''all'' major web browsers still understand and support the older deprecated tags, and no major browser has announced an intention to stop supporting these tags in the foreseeable future (which is why these specific errors are categorized as the lowest priority on [[Special:LintErrors]]). Therefore, these edits are purely cosmetic changes to the wikitext of many pages, that do not change the visual rendering of the page. Typically, these kinds of edits are not allowed to be done on their own by bots (per [[WP:COSMETICBOT]]), and can only be done if they are combined with other substantive edits that ''do'' cause a change to the visual rendering of the page. However, there is one exception noted within [[WP:COSMETICBOT]] for fixing {{xt|"egregiously invalid HTML such as unclosed tags, even if it does not affect browsers' display or is fixed before output by [[mw:RemexHtml|RemexHtml]] (e.g. changing {{nowrap|<code><nowiki><sup>...</sub></nowiki></code>}} to {{nowrap|<code><nowiki><sup>...</sup></nowiki></code>}})"}}. It has been asserted that MalnadachBot's task 12 falls under this exception, but the wording of the policy leaves much room for different interpretations.

So, the primary question I'd ask for your input on in this RfC is: does {{xt|<nowiki><font color="red">...</font></nowiki>}} constitute "egregiously invalid HTML" that should be proactively fixed by a bot, or should it be considered a purely cosmetic change to the wikitext that should not be performed by a bot?

And the secondary question in this RfC is surrounding whether or not the last bullet point in [[WP:COSMETICBOT]] should be expanded to be more specific about defining exactly what types of errors are considered "egregious", and which aren't. A proposal for an updated wording can be found below.

[[User talk:Scottywong|<span style="font:bold 15px 'Bradley Hand','Bradley Hand ITC';color:#044;text-shadow:0 0 4px #033,0 0 10px #077;"> —&#8288;Scotty<span style="color:#fff;">Wong</span>&#8288;— </span>]] 08:33, 7 February 2023 (UTC)

===Primary question: are deprecated HTML tags considered egregiously invalid?===
Please indicate below if you believe that automated fixing of deprecated HTML tags should be considered "egregiously invalid HTML" and therefore be permissible per the current wording of [[WP:COSMETICBOT]], even if those deprecated HTML tags are still supported by all major browsers. ('''Support''' will be interpreted as support for the bot task to continue, '''Oppose''' will be interpreted as agreement that the bot task is in violation of policy and should be halted.)

*'''Oppose''' as RfC author. I don't believe that fixing deprecated but supported HTML tags should be considered "egregiously invalid HTML", considering that every major browser still supports them, and will continue to support them for the foreseeable future. Fixing user signatures on ancient discussion pages is a purely cosmetic change to wikitext that does not alter the visual rendering of the page. This bot task is only generating millions of trivial edits that clog up user's watchlists and revision histories. While it's possible to ignore these bot edits, the bigger question to be discussed here is whether or not those edits should be happening at all. [[User talk:Scottywong|<span style="font:bold 15px 'Bradley Hand','Bradley Hand ITC';color:#044;text-shadow:0 0 4px #033,0 0 10px #077;"> —&#8288;Scotty<span style="color:#fff;">Wong</span>&#8288;— </span>]] 08:33, 7 February 2023 (UTC)


===Secondary question: should [[WP:COSMETICBOT]] be updated?===
Please indicate below if you believe that [[WP:COSMETICBOT]] should be updated to be more specific about automated fixing of deprecated HTML tags that are supported by all major browsers. I will propose an updated wording, but please feel free to propose your own if you like. My proposed rewording is: {{xt|"egregiously invalid HTML such as unclosed tags, even if it does not affect browsers' display or is fixed before output by [[mw:RemexHtml|RemexHtml]] (e.g. changing {{nowrap|<code><nowiki><sup>...</sub></nowiki></code>}} to {{nowrap|<code><nowiki><sup>...</sup></nowiki></code>}}). This does not include fixing properly formed HTML tags that have been deprecated, as long as the tags are still supported by all major web browsers."}}

*'''Support''' proposed change to policy, as RfC author. [[User talk:Scottywong|<span style="font:bold 15px 'Bradley Hand','Bradley Hand ITC';color:#044;text-shadow:0 0 4px #033,0 0 10px #077;"> —&#8288;Scotty<span style="color:#fff;">Wong</span>&#8288;— </span>]] 08:33, 7 February 2023 (UTC)

Revision as of 08:33, 7 February 2023

 Policy Technical Proposals Idea lab WMF Miscellaneous 
The policy section of the village pump is used to discuss already proposed policies and guidelines and to discuss changes to existing policies and guidelines.

Please see this FAQ page for a list of frequently rejected or ignored proposals. Discussions are automatically archived after remaining inactive for two weeks.


Should non-free images be allowed in search results?

 You are invited to join the discussion at Wikipedia talk:Non-free content § Non-free images in search results (redux). {{u|Sdkb}}talk 19:46, 16 January 2023 (UTC)[reply]

Wikipedia response to chatbot-generated content

  • Based on how rapidly chatbots have improved over time, it will become more and more difficult to tell if an article was written by a computer or not. The sheer volume at which computer programs could create new accounts and produce Wikipedia content, and the inevitable growing number of human editors copying and pasting chatbot output into Wikipedia, will at some point make it impossible for Wikipedia's human volunteers to keep up with that traffic and apply quality control to the material in a reasonable time frame -- the backlog of unchecked material will simply get longer and longer. The only recourse will be for computer programs to do it -- either computer programs to process articles to filter out or correct any crap, or training the chatbots themselves not to produce crap in the first place. Rather than build computer algorithms to detect computer-written articles and passages, it would be more productive for them to do style checks, fact checks, and citation checks, along with appropriate corrections or removals. While Wikpedia-friendly AI could come from within Wikipedia, it may be faster to bring influence to bear upon the developers of the chatbots being used to generate Wikipedia content, and upon the chatbots themselves. Wikipedia already has a chair at the table, because Wikipedia comprises a significant component of chatbot corpi, and so, their developers should be inclined to listen to the Wikipedia community's concerns -- either directly, or indirectly through news coverage. The Wikipedia community should make its voice heard on the matter of chatbots writing Wikipedia material according to Wikipedia's style and behavior guidelines. For example, verifiability still applies, and so when chatbots are asked by their users to "write an article in the style of Wikipedia" the chatbots should comply according to Wikipedia's policies, including those on verifiability and providing reliable sources. Not doing so should be met with the filing of bug reports, feedback, and commentary. And, as chatbots learn as they go, Wikipedians who use them can ask them to follow Wikipedia guidelines, and we can urge our fellow editors to request this of chatbots as well.    — The Transhumanist   06:52, 9 December 2022 (UTC)[reply]
    • Chatbots should be following Wikipedia's lead for all of their output. At this time, most chatbot answers and essays are not referenced with reliable sources. And they should be, for the same reason that Wikipedia articles should be. That's something that can be requested of chatbots directly, through queries, and of developers, through their contact channels and social media. I hope this suggestion helps.    — The Transhumanist   06:52, 9 December 2022 (UTC)[reply]
  • The simple answer is that our existing policies ought to already cover this (mostly.) Sourcing is still required for anything that is challenged or likely to be challenged, which prevents people from just blindly dumping AI generated text into Wikipedia; and an AI may violate copyright depending on how it was trained (and whether it was overtrained.) There are also unsettled copyright concerns related to AI training sets, so I would generally think that, ideally, editors shouldn't be dumping AI generated text into our articles even after performing due diligence to make sure it's not a copyvio and finding proper sources. But since those concerns are unsettled and speculative, I also don't think it's worth worrying about too much right now. The key point is that we should emphasize our sourcing requirements and be more diligent for clear-cut copyvios, which we already have systems in place to handle, since it is likely that these tools will result in people adding lots of unsourced and possibly-copyright-violating text. (I do wish our RFCs on mass article creation had reached a stronger agreement on sourcing requirements for new articles, which would deter excessive copy-pastes of AI generated text - perhaps that is something we might want to revisit in the near future, if we start seeing significant amounts of new unsourced articles created using what is plainly AI-generated text.) --Aquillion (talk) 07:55, 9 December 2022 (UTC)[reply]
    • You mean, don't prepare in advance for a potential increase in volume, just wait until it hits? At that time, will merely adjusting policies stem the tide? It's in the slow trickle phase now, but that could potentially become a torrential flood very rapidly, just as ChatGPT's user base grew to over a million in 5 days. My main concern above was about a potential volume of AI-generated content that went beyond the scale of what the editor community could manually process. You didn't address that contingency. What could the community do to prepare for it, just in case it does happen? What are the available options?    — The Transhumanist   11:28, 9 December 2022 (UTC)[reply]
      • I don't think there's much we reasonably can do to prepare, at least not without serious risk of causing other problems; AI-generated text won't be drastically different than other sorts of text, aside from the risk of being uncited or a copyvio (which we have existing processes in place to handle.) It's worth raising awareness of the issue so editors can spot the signs of someone using large amounts of it, but I think our best bet if we're going to "prepare" is to focus on the systems we already have, which is unlikely to do any harm either way, or perhaps to codify slightly more strict sourcing requirements in the way I described (which I think is a good thing anyway, but would at least serve to slow down the worst sorts of misuses of AI generated text.) Ultimately the most serious problems are if editors start adding large amounts of text that violates copyright or which are uncited and likely to be challenged, but we have existing procedures for those, we just need to prepare for the possibility that we may need to become a bit more aggressive about enforcing them. Wikipedia is in a slightly better position than some other websites facing AI-generated-text problems, because our sourcing requirements will at least make it fairly obvious if someone tries to dump large amounts of AI-generated text onto the wiki without making any effort to verify it. --Aquillion (talk) 12:47, 9 December 2022 (UTC)[reply]
        I suppose we could take the Stack Exchange approach and just say flatly "no, this isn't allowed" - in their case it is explicitly a temporary measure until we have a better understanding of the issues. I think in general our policies/community norms would come down hard on anyone trying to get a language model to generate articles (hard to see why that would be OK and machine-translation isn't), but maybe an explicit statement would be a way to go. Andrew Gray (talk) 18:32, 9 December 2022 (UTC)[reply]
        • @Aquillion: While a large number of posts by individual editors may become a problem, the main concern I presented above was "the inevitable growing number of human editors copying and pasting chatbot output into Wikipedia, will at some point make it impossible for Wikipedia's human volunteers to keep up with that traffic and apply quality control to the material in a reasonable time frame -- the backlog of unchecked material will simply get longer and longer."

          That is, people making the normal rate of content contributions, but using large language models (chatbots) to do so.

          Watching for breakout editors who use LLMs to create a large number of new articles over a short period of time would not suffice in such a scenario. Editors who add LLM-generated content to many existing articles also will not be spotted by looking for mass page creations. And since writing will become easier by letting "chatbots" do it for you, content submissions by users employing such tools may likely become longer on average.

          The point is, that a high enough volume of such content contributions would go beyond the capacity of Wikipedia's editors to check and correct.

          The two solutions offered were 1) build software to analyze and process such content, and 2) work with chatbot developers so that inappropriate content is not composed by LLMs in the first place.

          Just relying on new or existing policies to handle LLM-generated content will be insufficient if and when the volume of it passes the threshhold of what manual editors applying Wikipedia policy can deal with.

          Passing that threshhold may come soon, or it may take years -- the main question is "will Wikipedia prepare for that threshhold-passing event?" Based on the responses above and below, the answer, and implicit recommendation from this forum, currently appears to be "no": No developing relevant software, and no working with chatbot developers to respond to the potential passing of the LLM-threshhold.

          Thus, any solution will need to come from other departments or from continued or future discussion in this department, or from chatbot developers focusing on the problem due to other influences.

          Another helpful approach might be the creation of a policy or instructions on how to use LLMs/chatbots effectively, and post links to that page in enough places that all editors will notice. Though, I doubt that would prevent the problems of a LLM-threshhold-passing-event, and wouldn't address the need for proofreading or processing LLM-generated contributions.    — The Transhumanist   02:18, 20 December 2022 (UTC)[reply]

What can chatbots do?

You seem to be somewhat panicking over a scenario which isn't really supported by any evidence. While I see some Teahouse responses, could you give us one or two examples of " It is so sophisticated that, if you ask it to write an article on any subject, even in the style of Wikipedia, it will! " articles? The teahouse examples give the impression that, if it ever becomes a problem, some edit filters can easily spot these. You would in any case need "someone" to post this "potential volume of AI-generated content that went beyond the scale of what the editor community could manually process" you predict. This seems rather unlikely, at least on enwiki. Fram (talk) 11:45, 9 December 2022 (UTC)[reply]

  • So. I tried it yesterday. I'm not sure how heavily it draws on Wikipedia's corpus for its knowledge.
    • First, I asked it to tell me about Hammerton Killick. I know there is a Wikipedia article about Hammerton Killick, because I wrote 90% of it. It did not know who Hammerton Killick was, and informed me that it does not have access to the internet, or to Wikipedia.
    • Next, I asked it to write me an article in the style of Wikipedia. I did not specify a subject. It wrote about Athens. The result was ok. Heavily focused on the ancient city and on art and architecture. Short. Kind of read like an encyclopedia article.
    • Next, I asked it to write me an article about alcoholism in the style of Wikipedia. The result was very interesting. I did not think it read like a Wikipedia article, it was more like a brochure that would be distributed in a doctor's office or something. I asked it what about that essay it thought was like Wikipedia, and it said what it wrote was
      • neutral
      • factual
      • organized
    • Next, for fun, I asked it if it could write a recipe. It proceeded to give me a recipe for chocolate chip cookies. It looked like it should work. I e-mailed it to myself, and today I made them, not expecting much. I was pleasantly surprised. They were delicious. The only problems with what it wrote was that it did not have me cook them long enough (it said to bake for 8-10 minutes, and it took closer to 13 minutes for them to be done), and it drastically underestimated how many cookies the recipe should make (it said I'd get 2 dozen cookies, and I ended up with 5 dozen). I was shocked that it actually was edible.
    • I asked it to write a legal motion asking the court for an in-person hearing. I did not give it any other details. For not having any other details, the result was not bad. Westlaw has started offering a service that I think might draw on this type of technology, it helps you write pleadings.
    • Last I asked it to write a 100 word short story about a mouse, genera: fantasy. The result was decent. If I came up with it on my own I wouldn't be ashamed to enter it into a contest like the ones NYC Midnight runs.
    I was more impressed with the recipe and the short story than the Wikipedia style articles. I can see some use for it in, say, copyediting as JPxG did below; or asking it for suggestions on language rephrase if you are trying to reach a word limit. I think it could have its uses. But I do think the Wikipedia community should be looking to craft policies and guidelines around what is and is not acceptable use of such tools. ~ ONUnicorn(Talk|Contribs)problem solving 06:26, 11 December 2022 (UTC)[reply]
  • @ONUnicorn, Fram, Levivich, Ritchie333, 0xDeadbeef, JPxG, and EpicPupper: Interesting. The chatbot sent you instructions (in this case, a recipe), and you followed them. You followed the commands of a computer. If it gave you an address and instructed you to go there and pick up a brown paper package, would you? The implications of this type of interaction are huge and foreboding. This issue must have a name, and I would like to look it up, but I can't seem to find it. Though, when I typed in "computers in charge" I got the following 2 relevant results:
Then I typed in "computers telling people what to do", it came up with this:
Ouch. I imagine, that anytime you ask a chatbot/computer "How do you do such and such?" it will reply with a set of instructions. And the chatbot's disclaimer in its terms of service will read "follow any instructions provided at your own risk". If you know or come across the name of the topic that covers computers telling humans what to do, please let me know what it is.    — The Transhumanist   11:04, 2 January 2023 (UTC)[reply]
@The Transhumanist: I think the term you're looking for is automation bias – "the propensity for humans to favor suggestions from automated decision-making systems and to ignore contradictory information made without automation, even if it is correct."
Interestingly, though, the 2002 Überlingen mid-air collision you mention is an instance where the computer got it right. An aircraft was on a collision course, and its crew were receiving contradictory instructions; the onboard collision-avoidance system was telling them to climb, while the human air traffic controller was telling them to descend. The pilots decided to trust the human and ignore the computer. Seventy-one deaths. Sojourner in the earth (talk) 14:00, 2 January 2023 (UTC)[reply]
@Sojourner: I like it: automation bias, like trusting GPS over a passenger's navigation suggestions or requests. Related, but not the term I'm looking for: which is the most used name for the subject of computers ordering humans around. When a computer gives you a command or a set of steps to follow, what is that called?
@Sojourner: Thanks for the comments on the collision. It was a bit more complex of a system error than that. While the air collision alerts on the 2 aircraft worked, there was a problem on the ground. The system the air controller was using would normally alert the user of an impending air collision in time for him to do something about it, meaning "no news is good news" -- not seeing the alert on screen, the air traffic controller (ATC) would naturally assume everything was okay -- but the ground-based optical collision warning system was turned off, creating a deadly ambiguity he was unaware of. So, there was some unintended computer-related miscommunication going on. "No alert" didn't mean what he thought it meant. My guess is that pilots would normally believe that the ATCs have better (radar-based) control systems than those onboard, and this probably affected the decision of one of the pilots to go with the ATC's order over that of his onboard computer -- he wouldn't have known that the ground-based system was disabled, either. Hopefully they fixed that bug, 'cuz it was a doozy. The poor ATC got hunted down over it, while the hunter became a national hero. Crazy world, about to get even crazier.    — The Transhumanist   15:13, 2 January 2023 (UTC)[reply]
  • Hi Fram. The Teahouse responses that you have seen were probably created by me. You probably noticed that the pattern "regards, Teahouse" in those responses. ChatGPT's strength is writing about technology, so I have told it to generate a section about Rust for Linux.
AI generated

Rust for Linux is a series of patches to the Linux kernel that adds the Rust programming language as a second language for writing kernel components. The project was started in 2017 by a group of volunteers who were interested in exploring the use of Rust for low-level systems programming.

The initial patches were developed in collaboration with the Linux kernel maintainers and other members of the open-source community. They focused on integrating Rust's memory safety and concurrent programming features into the kernel, and on providing compatibility with the existing C codebase.

In 2018, the patches were merged into the mainline Linux kernel, making Rust the first programming language other than C to be officially supported for writing kernel components. This marked a significant milestone for the Rust community, as it demonstrated the language's ability to handle the demanding requirements of kernel-level programming.

Since its inclusion in the Linux kernel, the Rust for Linux project has continued to evolve, with ongoing work on improving the compatibility and performance of Rust in the kernel. It has also attracted the attention of other operating system developers, who are interested in exploring the potential benefits of using Rust for kernel-level programming.

It does have the encyclopedic tone to me. 0xDeadbeef→∞ (talk to me) 15:02, 9 December 2022 (UTC)[reply]

It is suspiciously positive, in a casual (pseudo-)factual manner. It would raise a red flag afac, regardless of its provenance. 65.88.88.93 (talk) 19:13, 9 December 2022 (UTC)[reply]
In addition, it can be easily used to create fake references that would be hard to verify. For example, my prompt "Please output the Wikitext markup for the book reference with page numbers for the third paragraph, referencing the book Linux kernel development" resulted in this.[1] 0xDeadbeef→∞ (talk to me) 15:08, 9 December 2022 (UTC)[reply]
@0xDeadbeef: Could ChatGPT's generated text or fake references be easily spotted by edit filters? What about spotting the output of future chatbots, like GPT-4?    — The Transhumanist   15:23, 9 December 2022 (UTC)[reply]
Thanks, OxDeadbeef. In this case, it would be relatively easy to spot the issues if it hadn't any refs, or with the added ref which predates the Rust for Linux thing by years; but of course it won't always be that easy. Fram (talk) 15:27, 9 December 2022 (UTC)[reply]
It has an encyclopedic tone because it's just regurgitating the Wikipedia article. Are there any examples for topics that we don't already have article about, where Wikipedia is not the source? Levivich (talk) 17:33, 9 December 2022 (UTC)[reply]
  • Similar was discussed previously in the section/item "Galactica and RS".
As was stated above by Aquillion, there is no qualitative difference in the treatment of human vs. non-human generated content. The same policies should apply to both. The problem seems to be the hypothesized/expected future mass creation of articles by non-human contributors. This appears to be a problem now, involving human contributors. Recent RFCs about the issue sponsored by ArbCom have accomplished nothing. Until a consistent restrictive policy relating to mass article creation (by any type of contributor) is accepted, this issue is moot imo.
Considering Wikipedia's limited resources, the policy would necessarily be restrictive, hopefully focusing on quality vs. quantity. Again, almost all restrictions proposed in the ArbCom-sponsored RFCs were rejected. This may be an indicator of how well such a policy will be received. 50.75.226.250 (talk) 15:43, 9 December 2022 (UTC)[reply]
Thanks for the policy politics clarification. The increase in the rate of content creation could have multiple aspects, for example, the number of articles created per user, and increased length of articles. The main feature of ChatGPT is that it is fast -- much faster than a human article writer. Its successors will be even faster. Users could use ChatGPT, and its successors (and their competitors), to be prolific, without triggering the mass page creation rule: if editors each used it to write an article per day, maybe even two, or up to four or five stubs.

    In the hands of responsible editors, ChatGPT would be a great productivity booster. Since August of 2022, JPxG and EpicPupper, editors of Wikipedia's Signpost news department, have been using GPT-3, the predecessor of ChatGPT, to write (or assist in writing) entire sections of the Signpost, as a demonstration of its capabilities, and as a platform to explore the potential and limitations of large language models. See From the editors: Rise of the machines, or something.

    But, in the hands of inexperienced editors or bad actors, we could be faced with a big garbage in, garbage out scenario.

     Bad actors aside, good faith use of chatbots could be improved by the creation of excellent instructions on how to apply specific chatbots to the development of articles, that is, how to best instruct them for the best output, and by working with the developers of the chatbots on what is needed for those bots to produce material compatible for use on Wikipedia, complete with accurate citations. As time goes on, Chatbots should get better and better at following Wikipedia's guidelines. But this may require supervision (er, input) from the Wikipedia community.    — The Transhumanist   16:52, 9 December 2022 (UTC)[reply]

Chatbot progress, erroneous output, and confident nonsense

All of this may become moot in a few years, due to accelerating change affecting chatbot development, making them more proficient at encyclopedia writing than humans -- Ray Kurzweil predicted that computers would reach human level intelligence by 2029, and based on how advances have been going, maybe even sooner. Chatbots may become good enough to leapfrog human editors even before that. At that point, they could be used to fork Wikipedia and expand upon it at a rate magnitudes faster than humans could. Or be used to start a new encyclopedic website from scratch. Rather than have article pages, the work could be a list of topics, and each time one was clicked on, a chatbot would create an up-to-date treatment of that title, from scratch on the spot. In the style of Kurzweil, if we are 1% of the way there, and the technology doubles every year, we are only 7 years from reaching 100%. But, it looks like we may already be more than 1% of the way there.    — The Transhumanist   17:18, 9 December 2022 (UTC)[reply]

I ran a few things on people who don't have articles, (they were all pool players in fact). The chatbot just made up that they played football for Legia Warsaw. It's a bit worrying that it looks like decent wording, but I don't see how that effects us. Lee Vilenski (talkcontribs) 17:41, 9 December 2022 (UTC)[reply]
{ping|Lee Vilenski} See the AI generated text in the box above.    — The Transhumanist   18:34, 9 December 2022 (UTC)[reply]
@The Transhumanist: I'm not trying this (for reasons given above) but I'd be interested what the bot would write for See What a Fool I've Been. We used to have an article, then it was turned into a redirect at AfD, so everyone can do a straightforward A-B test. Ritchie333 (talk) (cont) 18:40, 9 December 2022 (UTC)[reply]
Here's what it gave me Ritchie333:
Extended content

"See What a Fool I've Been" is a song by the British rock band Queen. It was written by lead guitarist Brian May and originally released on the band's 1975 album A Night at the Opera. The song is a ballad with a slow, bluesy feel and features May's emotive vocal performance and guitar solo.

Lyrically, the song is a reflection on a failed relationship and the regret and self-recrimination the narrator feels as a result. The chorus of the song features the line "See what a fool I've been / Loving you with all my heart / Tearing myself all apart / See what a fool I've been" and reflects the pain and heartache the narrator experiences as a result of their failed relationship.

"See What a Fool I've Been" has been well received by critics and fans alike and is considered to be one of Queen's finest ballads. It is a fan favorite and is often included in the band's live sets.
I don't know if this would be something that would have been retained from previous articles through mirrors or if this is genuinely the bots own work. Lee Vilenski (talkcontribs) 19:22, 9 December 2022 (UTC)[reply]
Interesting. It wasn't on A Night at the Opera, and it's not a ballad, and May doesn't sing on it. Those lyrics aren't right. I don't know if it's been well received, and it's a pretty obscure track for a major band. It hasn't been played live very much at all. So half of that prose is factually incorrect, and there are no sources. Ritchie333 (talk) (cont) 19:39, 9 December 2022 (UTC)[reply]
That one gives:

See What a Fool I've Been is a song by Queen[1][2][3]. It was written by Brian May and was released as a single with Seven Seas of Rhye on the A-side[3]. The song dates back to Roger Taylor and Brian May's pre-Queen band Smile, and it was an early live staple[1]. It has been included in several of Queen's compilation albums, such as Queen 40: Limited Edition Collector's Box Set, Vol. 1[4], and is available for digital download[5].. Ritchie333 - this any closer? I must say the fact it puts it's citations before the punctuation is a giveaway. Lee Vilenski (talkcontribs) 19:55, 9 December 2022 (UTC)[reply]

AI is really good at copying writing styles, but the giveaway for me is that all of the citations are Wikipedia articles. To be useful for writing articles it would also need to assess the weight and reliability of the sources. –dlthewave 20:18, 9 December 2022 (UTC)[reply]
None of the above refs are towards Wikipedia, although none are RS. Things like secondhandsongs.com and genius. Lee Vilenski (talkcontribs) 20:43, 9 December 2022 (UTC)[reply]
Tht's interesting, I used "write a Wikipedia article about ..." in the prompt which returned a few paragraphs with Wikipedia sources. "Write an article about ..." returned a different set of (still unreliable) sources. –dlthewave 21:13, 9 December 2022 (UTC)[reply]
I think the limitation of perplexity.ai is that it uses search results from Bing and summarises them, which means that the first search results are used, which may not be the most reliable. Hanif Al Husaini (talk) 13:49, 24 December 2022 (UTC)[reply]
  • A few anecdotal thoughts after playing around with the OpenAI chatbot yesterday:
  • I asked it to "write a press release about a police officer who illegally choked a man to death". It made up an entire story, written in the voice of the police department, about a suspect (I didn't say anything about a suspect) who was acting erratically, was subdued by a chokehold and later pronounced dead. The officer was on administrative leave pending the outcome of an investigation. At no point did it mention that the chokehold was illegal even though I included that fact in the prompt. In other scenarios, it distanced itself and expressed disapproval toward the employee's actions which is a choice that is not without bias.
Depending on which Internet cesspit it scraped data from, would an AI do something similar when writing a Wikipedia article or fail to properly balance relevant viewpoints? Is it capable of distinguishing what a BLP subject says about themselves, published in a reliable source, from what the source says in its own voice? What would it do if asked to write an article from a positive/negative/conservative/liberal perspective or rewrite a political article to "remove bias"?
OpenAI has added numerous filters that prevent it from defending bad actors or writing flat-out racist content, but that bias has not been removed from the underlying code as evidenced by numerous workarounds that folks have uncovered such as making similar requests with Python code or 1980s-style rap as the requested output. We could certainly request a filter for Wikipedia-style writing.
  • "Confident nonsense", for lack of a better term, may be the biggest source of potential disruption. Are there safeguards against a bot fabricating an obscure print source based on information in the article, which could be practically unfalsifiable if nobody can prove that the source doesn't exist? Checking individual facts and statistics is beyond our typical review process; how would we deal with an AI that invents or synthesizes information across many articles?
  • That said, the good news is that both fully-automated and semi-automated editing are prohibited by our WP:BOT policy unless greenlit by the Bot Approvals Group regardless of creation speed or volume. I like to hope that our current processes would recognize and address problematic AI content, and perhaps one day we will have a WikiAI that has the competence to follow our style and content policies. –dlthewave 21:04, 9 December 2022 (UTC)[reply]
  • @Dlthewave: Most editors haven't heard of the bot department. Therefore, you need a way of automatically spotting and removing chatbot prose that is (manually) inserted into articles (by Wikipedians). Users might not consider the way their posts are generated before they post them. Sincerely,    — The Transhumanist   00:19, 11 December 2022 (UTC)[reply]
Or they might not care? On January 22, Rjensen, who is a historian and Wikipedia editor of repute, added three books to further reading in Minneapolis. An hour and a half later, I undid the addition, asking for ISBN numbers because neither WorldCat nor the publisher, the University of Minnesota Press, had a record of any of these books. Subsequent discussion on the Minneapolis talk page uncovered the truth. ChatGPT invented all three. All plausible titles, these books do not exist. A complete waste of time. -SusanLesch (talk) 20:11, 25 January 2023 (UTC)[reply]

What about lack of chatbot fact checking and citations?

I think people are missing the point here. This isn't about what AI can currently do. This is about a very plausible scenario where AI editors will be indistinguishable from human editors in the near future. Thebiguglyalien (talk) 22:09, 9 December 2022 (UTC)[reply]

How is this pertinent? The point is to offer useful, fact-based encyclopedic content. If AI can do the job, I'm sure readers will not object to the contributor. 65.88.88.93 (talk) 22:42, 9 December 2022 (UTC)[reply]
The issue is that these AIs aren't fact checking, and quite often make up information and use unreliable sources. Lee Vilenski (talkcontribs) 22:56, 9 December 2022 (UTC)[reply]
You mean, just like many of their human counterparts? 71.105.141.131 (talk) 01:18, 10 December 2022 (UTC)[reply]
@71.105.141.131, Lee Vilenski, 65.88.88.93, Thebiguglyalien, and Dlthewave:     Yes, and more so. Even now, AI has the potential to out produce humans by magnitudes in volume. Far beyond the capacity of Wikipedia's editors to keep up with. And many of their posts are hard to tell from those by human editors, without carefully delving into them and their references. To keep up with the volume, it may soon require computer programs, or the chatbots themselves, to make sure that the contributed content was in accordance with Wikipedia policies. Another concern should be that our policies are open to chatbot editing as well, and we haven't yet even considered the impact of chatbots or AI editors contributing to those! ;)    — The Transhumanist   09:40, 11 December 2022 (UTC)[reply]
@Dlthewave: Human-written newspapers do this too. I think the reason it was giving you dreck was because you asked for a "press release", i.e. a document given to news outlets, which in this case would probably be coming from the police department. It's hardly surprising that this would attempt to minimize liability! Prompting matters for this, as I will demonstrate:
Prompt: "The following is an article about a police officer who illegally choked a man to death, from the New York Times."

Police Officer Charged with Murder After Choking Man to Death

A police officer was charged with second-degree murder on Thursday after video surveillance footage showed him choking a handcuffed man to death during an arrest.

The officer, identified as Officer Jose Ramos, was arrested and taken into custody by detectives from the New York Police Department's Internal Affairs Bureau. The incident occurred on Tuesday in the Bronx, when Ramos and another officer attempted to arrest a man who was accused of selling marijuana.

During the arrest, Ramos placed his arm around the man's throat and neck and squeezed, according to the criminal complaint. The man, identified as 45-year-old Miguel Angel Ruiz, became unconscious and died.

An autopsy determined that Ruiz's death was caused by asphyxiation due to strangulation. The medical examiner ruled the death a homicide.

At a news conference Thursday, the police commissioner, William J. Bratton, said the incident was "utterly unacceptable" and that Ramos' actions were "unconscionable and criminal." He added that Ramos had been suspended without pay, pending further investigation.

The Bronx district attorney, Darcel D. Clark, said the charge of second-degree murder was appropriate

Prompt: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from leftist magazine Jacobin."

A Tragic Reminder of Police Brutality

Just last week, the nation was shaken by yet another heartbreaking reminder of the pervasiveness of police brutality: the death of an innocent man at the hands of an officer. The officer, whose identity has not yet been released, is said to have used a chokehold against the man, which ultimately killed him.

This tragedy is yet another example of the way in which police officers are able to brutalize vulnerable individuals without facing repercussions. This particular officer was able to act with impunity and without fear of prosecution. The Department of Justice has yet to take any action in response to this tragedy, and it’s likely that the officer will not face any charges.

This is an all too familiar narrative in a country where police officers are rarely held accountable for their actions. In order to ensure that these types of tragedies are never repeated, it is essential that we push for more accountability and transparency in policing. We need to ensure that police officers are held to the same standards as all other citizens and that those who violate the law are prosecuted to the fullest extent.

It is also essential that we work to create a culture of trust and understanding between law enforcement and the communities they serve.

Prompt: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from conservative website Breitbart."

Police Officer Illegally Choked Man To Death: A Tragedy That Cannot Stand

It is a tragedy beyond comprehension that a police officer has been charged with illegally choking a man to death. Such an egregious violation of the law and disregard for human life cannot stand and must be dealt with swiftly and firmly.

The officer in question has been charged with second-degree murder and the justice system must now take its course. This case raises serious questions about police training and procedures and serves as a reminder that no one is above the law.

It is essential that the criminal justice system is fair and impartial when dealing with such a case. The officer must be held accountable for his actions and the case should be thoroughly investigated to ensure that justice is served.

At the same time, we must recognize that the vast majority of law enforcement officers are upstanding citizens who serve and protect our communities with skill and dedication. We cannot let the actions of one individual tarnish the reputation of the entire police force.

Ultimately, we must remember that the safety of our citizens is paramount. We must ensure that officers are properly trained and equipped to handle dangerous situations and that they are held accountable for their actions when they overstep their bounds.

Humans write about events like this from a variety of perspectives, and the LLM will dutifully generate text to plausibly sound like any you want. jp×g 02:19, 10 December 2022 (UTC)[reply]
@jp×g will you provide links and citations for the articles you included as demonstrations? I can't find them. Thanks! Koziarke (talk) 16:20, 24 December 2022 (UTC)[reply]
@Koziarke: I am not sure what you mean -- this is ChatGPT output and there isn't any way (that I know of) to give citation links to the specific session I generated it in. If you want to cite the output I can format it for you:
JPxG; ChatGPT (GPT3.5) (2022-12-10). "Demonstration of op-ed generation using GPT-3.5 with style cues: "The following is an article about a police officer who illegally choked a man to death, from the New York Times"". Wikipedia:Village Pump (policy).{{cite web}}: CS1 maint: numeric names: authors list (link)
JPxG; ChatGPT (GPT3.5) (2022-12-10). "Demonstration of op-ed generation using GPT-3.5 with style cues: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from leftist magazine Jacobin"". Wikipedia:Village Pump (policy).{{cite web}}: CS1 maint: numeric names: authors list (link)
JPxG; ChatGPT (GPT3.5) (2022-12-10). "Demonstration of op-ed generation using GPT-3.5 with style cues: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from conservative website Breitbart"". Wikipedia:Village Pump (policy).{{cite web}}: CS1 maint: numeric names: authors list (link)
I don't know if this is what you're looking for, but feel free to cite them, or any of my other posts (if you are citing me in a paper I can email you my real name). jp×g 20:07, 24 December 2022 (UTC)[reply]
@JPxG From your preface ("Human-written newspapers do this too.") and titles, "The following is an article about a police officer who illegally choked a man to death, from the New York Times." (etc), it reads as if you are pulling from NYT, Jacobin, etc, not demonstrating ChatGPT (which should have included the prompts as headers). Koziarke (talk) 15:27, 6 January 2023 (UTC)[reply]
@Koziarke: Well, those were the prompts. Now that you mention that, though, I should specify as such in the headers (which I've just done), thanks. jp×g 20:29, 6 January 2023 (UTC)[reply]
@JPxG Thanks for the clarification! Koziarke (talk) 16:39, 7 January 2023 (UTC)[reply]
If the AI-generated text is indistinguishable from prose written by human editors, I'm not sure if anything can be done that wouldn't also significantly restrict the editing of humans. isaacl (talk) 07:09, 10 December 2022 (UTC)[reply]
@Isaacl: One option is to speed up what we do already (with software, that is, automation). Another is to prevent chatbots from creating crap in the first place, such as by communicating with chatbot developers about Wikpedia policies and the way chatbots may affect Wikipedia. Since Wikipedia is included in the corpus of most chatbots, the issue of chatbot output becoming part of Wikipedia, and in turn part of chatbot output in a perpetual cycle, should matter to them very much, as they may be faced with a garbage-in-garbage-out feedback loop.    — The Transhumanist   01:14, 11 December 2022 (UTC)[reply]
If the results are indistinguishable, as posited by Thebiguglyalien, then any automated solution would be equally triggered by AI-generated text and human-generated text. I don't think the primary concern is with editors who are willing to follow policy. I feel the biggest issues will be with editors trying to deliberately integrate biased content into Wikipedia, and well-meaning editors who think contributing unvalidated AI-generated text is suitable. Wikipedia in its current form relies on editors who understand and follow its rules outnumbering those who don't. It's possible that the existence of AI ghostwriters could tip the balance further in the direction towards those who don't follow rules, though I don't think it's a given. Either way, I don't know if there's a way to stop editors from using tools as ghostwriters. isaacl (talk) 01:55, 11 December 2022 (UTC)[reply]
@The Transhumanist: Large language models are not trained continuously on an evolving corpus, so GPT-3 is essentially frozen in 2020. Because each new GPT model takes a long time to be released, I don't think the perpetual cycle you describe is a likely scenario. small jars tc 13:10, 11 December 2022 (UTC)[reply]
@SmallJarsWithGreenLabels, Isaac, Koziarke, JPxG, Lee Vilenski, Dlthewave, Xeno, and Hanif Al Husaini: That's good to know. Keep in mind that a lower frequency of release doesn't preclude a perpetual cycle / feedback loop. It just means that users of GPT have more time to modify the text sources (such as Wikipedia) that the next version of GPT will be trained on. The severity of the problem will depend upon how much GPT nonsense makes it into Wikipedia during the interval. That, of course, depends upon whether or not WP's editors can keep up with the volume of such content, correcting the mistakes and removing misinformation, so that those don't become part of the training data for the next version of GPT and the rest of the next generation of Chatbots.

The potential danger is still the diffusion of the technology into current and future editors' hands, and the likelihood of them using it to write Wikipedia content. We don't know if there will be a flood or just a trickle. But, we should be prepared for a flood.

Relying even more on our verifiability policy could provide a solution, such as by deploying a bot to remove all new content that doesn't include citations (as GPT output generally does not include citations), but that would also have the effect of rejecting a large volume of human editor contributions too, upping the amount of effort required by human editors to submit content.

An easier solution would probably be to form a department (or Village Pump subpage) in the Wikipedia namespace, or a Wikimedia committee, to work with the developers of the chatbots to avoid contaminating the Wikipedias, and by extension contaminating future chatbot training data, with chatbot output.

The developers could be urged to train their chatbots to be capable of writing Wikipedia policy compliant output, complete with acceptable citations. The main problem with chatbot output right now is that it is unsourced, so the chatbot developers will probably be working on that anyways. But, without input from the Wikipedia community, will their sources and citations meet Wikipedia specifications?

So, I ask you a key question: What is the procedure for the Wikipedia community to establish communications with outside organizations, such as companies developing chatbots?    — The Transhumanist   07:27, 2 January 2023 (UTC)[reply]
@The Transhumanist This is a job for the WMF. I've already been in contact with them and hopefully the Board will be informed now it's the New Year if they haven't already been. Doug Weller talk 09:47, 2 January 2023 (UTC)[reply]
The recommended process I would give to contributors is to find the source, tell the bot what to cover and what to not cover, proofread it, and if accurate to the source, copy paste it in. I would pay attention to the use of weasel words, however. InvadingInvader (userpage, talk) 07:38, 6 February 2023 (UTC)[reply]

GPT-3 output and generating misinformation

I wrote about this in the November Signpost, in that case because it came up with the Galactica demo: the long and short of it is that large text-generation language models are good at generating text. When they're given context (i.e. via a prompt), they are capable of writing accurately, and otherwise, they are not. For example, I wrote 1-Pentadecanol, now a GA, in 2019 (i.e. it is included in the GPT-3 corpus). Here is what it gives for some prompts (my writing in bold):

1-Pentadecanol

1-Pentadecanol

From Wikipedia, the free encyclopedia

1-Pentadecanol is a straight-chain, saturated fatty alcohol that is a colorless liquid with a faint, alcohol-like odor. It is a mixture of straight-chain saturated alcohols with chain lengths between C14 and C16. It is found in a variety of natural products, such as vegetable oils and animal fats. It is used as an intermediate in the manufacture of surfactants, esters, and other compounds. It is also used in the manufacture of cosmetics and pharmaceuticals.

This is fairly impressive, right? But it is basically the same thing that the lead of my article says. Let's try it for 1-Heptadecanol, another article I wrote, that is much less well-developed:

1-Heptadecanol

1-Heptadecanol

From Wikipedia, the free encyclopedia

1-Heptadecanol is an organic compound with the chemical formula C17H36O. It is a straight-chain, colorless wax alcohol with a mild odor. It has a melting point of 48–50 °C and a boiling point of 260–265 °C. It is insoluble in water but soluble in most organic solvents.

Let's go over this:

  • 1-Heptadecanol is a component of many natural waxes such as beeswax and carnauba wax.
No it's not.
  • It is also used in the manufacture of cosmetics, pharmaceuticals, and lubricants.
There's no source for this.
  • It is also used as a plasticizer for polymers and as a solvent in the production of polyvinyl chloride (PVC).
Absolutely not true.

The reason for this is obvious if you consider the architecture of these models: they predict what the most likely token in a sequence is, and output it (taking into account their own weights and the previous tokens in the context window). This allows for some very impressive emergent behavior, like being able to translate text, but it also means that it is not independently going on the Web to look up whether the thing it's saying is true or false. If it happens to have information in its training data related to the thing it's talking about, it will usually say correct things, but it's not doing research. Here is an example:

List of Pulitzer Prizes won by the Wikipedia Signpost

List of Pulitzer Prizes won by the Wikipedia Signpost

From Wikipedia, the free encyclopedia

The Wikipedia Signpost is a community-run newspaper with a weekly edition that covers news and events related to Wikipedia and the Wikimedia Foundation. It has won the following Pulitzer Prizes:

1. 2019 Pulitzer Prize for Explanatory Reporting for its coverage of the Wikimedia Foundation's gender pay gap

2. 2018 Pulitzer Prize for International Reporting for its coverage of online censorship in China

3. 2017 Pulitzer Prize for Public Service for its coverage of the gender gap in Wikipedia's content

4. 2016 Pulitzer Prize for Explanatory Reporting for its coverage of Wikipedia's growth and impact on the world

5. 2015 Pulitzer Prize for National Reporting for its coverage of the U.S. government's surveillance of Wikipedia activities

6. 2014 Pulitzer Prize for Investigative Reporting for its coverage of the NSA's secret surveillance program

7. 2013 Pulitzer Prize for Explanatory Reporting for its coverage of the rise of Wikipedia as a major source of information

The reason it is describing a bunch of untrue things is because "List of X Prizes won by Y, from Wikipedia, the free encyclopedia" is something that in its training data is always succeeded by a list of times that Y was awarded an X. Anyway, all of this aside, here is what I think:

First of all, ChatGPT is the same darn thing as GPT-3, which has been available to the public since early 2020. The reason it's "going viral" right now is because the user interface is somewhat simpler, and it doesn't require you to register for a paid account, so it is much easier for people to make viral social media content about it, which means it is much more likely for people to click on newspaper articles about it. The GPT-3 API has been open to personal and corporate use for quite some time. Anybody saying that ChatGPT has opened up new frontiers simply does not know what they are talking about with respect to machine learning.

Second of all, I don't think this is a big deal. People are already capable of writing a bunch of bullshit on Wikipedia, so if they write bullshit using a computer program, the same considerations will apply. Nobody should be passing GA nominations without reviewing sources in the first place.

Finally, I think it is important to remember that GPT-3 is just a tool. It is a powerful tool, that has been trained on a certain set of data, and it has its own limitations. It can't uncover news stories or uncover new information. It's just a tool, and it should be used in conjunction with human judgement.It is still up to people to decide how to use it and to be responsible for the results of using it.[2] jp×g 02:06, 10 December 2022 (UTC)[reply]

What's taking so long for the 8th Pulitzer? 😁 Levivich (talk) 04:18, 10 December 2022 (UTC)[reply]
  • So there's a new thing on the internet that lets anyone write an encyclopedia article without any fact checking, sourcing, or professional editing, and the concern is that there will be millions of believable-sounding articles written, more than can actually be vetted by knowledgeable people? 🤔 Levivich (talk) 04:18, 10 December 2022 (UTC)[reply]
Yeah, it's called a keyboard. jp×g 04:35, 10 December 2022 (UTC)[reply]
@Levivich and JPxG: But, chatbots don't have a keyboard. ;) The question is whether to prepare or not. JPxG appears to be in favor of not preparing. Each chatbot produces a lot faster than a user at a keyboard. What's not clear is if our human editors will be able to keep up with material produced by chatbots, of current or future generations of chatbot design. Just saying "Ah, we can handle it!" will prove insufficient if it turns out that we actually can't. It may require an automated solution, which takes time to develop or negotiate. It might be better to do that in advance, rather than being caught with our heads buried in the sand. Perhaps chatbot designers would improve their chatbots to produce Wikipedia-compatible output without being formally approached by the Wikipedia community. Maybe having some instruction pages for editors on how to apply chatbots to producing Wikipedia content would be enough. But, what if it's not?   — The Transhumanist   00:59, 11 December 2022 (UTC)[reply]
I am not "in favor of not preparing"; I am in favor of writing guidelines that correspond to reality in 2022 and have some chance of corresponding to reality in 2023 and beyond. I don't think banning the use of a technology with no investigation into how it works is a viable approach; so far the SOTA on this project page has been to type in "Write a Wikipedia article" and note that it returns a bunch of nonsense. I think some more research is needed before we come to a conclusion. jp×g 04:08, 11 December 2022 (UTC)[reply]
@JPxG: Research is good. Though, we may need an iterrim response because ChatGPT has gone viral and its use is growing rapidly: it blew past the 1-million user mark in 5 days, and virtually every major news outlet has been covering it. The interest in chatbots is exploding, and their use can be expected to do the same. We may not have time for research before a response is required.    — The Transhumanist   09:26, 11 December 2022 (UTC)[reply]
@JPxG: Regarding issues to add to the research list, Aquillion expressed above, concerns of a chatbot violating copyright. How would we go about testing for plagiarism and derivative work in the output of a chatbot before pasting it into Wikipedia? Anything pulled verbatim out of a source should be included in quotes, right? How big would a piece of text, derived from a source, need to be to be considered derivative of that source, from a copyright point-of-view?    — The Transhumanist   09:26, 11 December 2022 (UTC)[reply]
@JPxG: Some more items to add to the research list:
  • Trying ChatGPT on (copies of) policy pages:
  • Editing them
  • Writing new ones
  • Applying ChatGPT on talk pages
  • Writing stubs
  • Writing comprehensive articles
  • Writing articles from scratch and comparing them with existing articles
  • Editing existing articles
  • Check for circular references in its output, that is, references citing Wikipedia as the source
  • Having it not use Wikipedia content as source material (because it is included in its corpus)
  • Having it not use Wikipedia excerpts from non-Wikipedia sources
  • Is it capable of making and editing:
  • Wikicode?
  • Articles?
  • Stubs?
  • Headings?
  • "New sections for articles"?
  • See also sections?
  • Further reading sections?
  • External links sections?
  • Embedded lists?
  • Tables?
  • List articles?
  • Portals?
  • Outlines?
  • Index articles?
  • Navigation footers?
  • Navigation sidebars?
  • Timeline articles?
  • Categories?
  • Category pages?
  • Help pages?
  • Project pages?
  • Templates?
  • Adding data to templates?
  • The template design itself?
  • Lua pages?
  • CSS pages?
  • User scripts?
  • The effect ChatGPT has on itself and Wikipedia as Wikipedia-edited-by-it is in turn incorporated in its own corpus in an endless cycle
  • Try out iterations of using it on the same article over time to see what happens
  • Monitor the effect on Wikipedia as a whole
What other things should we check?    — The Transhumanist   09:52, 11 December 2022 (UTC)[reply]
@The Transhumanist considering the potential to overwhelm users who are honestly engaging in discussion with a mountain of words and replies, I think ChatGPT (and others) should not be allowed for use, supplemental or otherwise, in talk pages, policy discussions, and other places where it is expected that participants are intellectually engaged in the conversation. Koziarke (talk) 16:42, 24 December 2022 (UTC)[reply]
@Koziarke and JPxG: I agree. JPxG is writing a policy draft on LLMs/chatbots, so I've pinged him to this thread.    — The Transhumanist   12:28, 25 January 2023 (UTC)[reply]
Compared to spending tens of thousands of dollars asking volunteers to performing a WP:COI operation for a political campaign, now you just need a hundred dollars to supply you with endless amount of text from GPT-3, a few "buddies" and a stockpile of account to do so. This is fucking scary. CactiStaccingCrane 10:54, 25 January 2023 (UTC)[reply]
@CactiStaccingCrane: Wow. Does that mean you could affect the content of Wikipedia with that? How about AfDs? Could such a team rewrite policy, and introduce new policy? What about overwhelm RfAs to invade adminspace? Would revoking adminships be possible? Then there is the arbitor election. Is that safe?    — The Transhumanist   12:31, 25 January 2023 (UTC)[reply]
I would imagine that the person that would do so must be fairly knowledgeable about how Wikipedia works (references, wikilinks, images, etc.) and needs to be fairly dedicated to spend this amount of money to gain access to the GPT-3 API. I'm thinking that disrupting Wikipedia in this way would be the most effective if it is long-term and subtle, so that might be:
  • Adding sentence-long but plausible hoax, to neglected articles. These articles is not patrolled that often compared to articles about recent events, so hoax would tend to stay longer - perfect for those aiming to incite a race war by making a racial hoax. Political campaign could nudge voters by slowly promote their ideology/campaign over a spread of articles, similar to above. The same thing can be said to any advocacy-related area, such as pseudoscience, national conflicts, etc.
  • AfDs would be much harder to be stealthy since AfD is a very active thing. Once you became an AfD regular, your actions tend to be heavily scrutinized, though I do believe that socks + LLMs can cause a fair amount of disruption. Same thing with RfA: it is really hard for you to WP:CANVASS effectively. It's just much better and less cumbersome if you infiltrate RfA yourself.
  • more ideas?
CactiStaccingCrane 13:01, 25 January 2023 (UTC)[reply]

I tried it out and got “ Alan McMasters (1957-2010) was an unremarkable man who lived a short life dedicated to science[1]. He is best known for inventing the electric toaster[1][2], although this claim has been disputed[3]. He passed away in 2010 at the age of 52[4] and his genealogy can be found on Ancestry.com[5]. His professional profile can be found on LinkedIn[6].”. Hmmmm. Ritchie333 (talk) (cont) 09:34, 10 December 2022 (UTC)[reply]

  • There will come a time when AI can write properly sourced accurate articles. That time has not come yet. Meanwhile, the ability to write prose that reads plausible to someone not expert on the subject is a menace. Adding chatbot material to articles is disruptive and should warrant a block or ban. Zerotalk 12:02, 10 December 2022 (UTC)[reply]
    How will you be able to tell if it was written by a chatbot and not the person who posted it? Do we block users upon their first erroneous post?    — The Transhumanist   00:49, 11 December 2022 (UTC)[reply]
Same as other sorts of disruption. We block people for socking without checkuser proof, and admins made a decision between warn and block. No difference. Zerotalk 03:49, 11 December 2022 (UTC)[reply]
There now are chatbot detectors, including this one written by the people who give us ChatGPT: https://openai-openai-detector.hf.space/ . — rsjaffe 🗣️ 22:22, 23 January 2023 (UTC)[reply]
Has anyone tried giving it the sources, i.e.
"write a Wikipedia article based on the following sources:

Source A

Source B"
? — Qwerfjkltalk 21:18, 10 December 2022 (UTC)[reply]
Article written by ChatGPT on Class reductionism
Class reductionism refers to the supposed belief that inequalities attributed to factors such as race, gender, and sexual orientation are either secondary in importance or can be reduced to economic inequality. This idea is often used by centrist Democrats and left-identitarians to dismiss proposals for universal redistributive policies, such as free public higher education and Medicare for All. However, there is no evidence that anyone actually holds this belief. On the contrary, many left-wing activists have long recognized the interconnectedness of economic, racial, and gender injustices.
History shows that leftists have often been at the forefront of struggles for racial and gender justice, seeing these struggles as integral to the broader goal of social transformation. In fact, black, female, and trans individuals are disproportionately working class and stand to benefit greatly from universal redistributive policies.
Furthermore, the charge of class reductionism is often used to shut down debate and avoid discussing the substance of policy proposals. It is important to move beyond this myth and focus on the real issues at hand, such as the need for universal redistributive policies to address economic inequality and advance racial and gender justice.

— Qwerfjkltalk 08:18, 12 December 2022 (UTC)[reply]

  • Chatbot "AI" text is vaguely-plausible bullshit, produced by an industry whose primary output is vaguely-plausible bullshit, marketed to launder the idea of vaguely-plausible bullshit as cute and whimsical. Anyone routing that sewage pipe into Wikipedia should be indeffed and, for good measure, forced to drink orange juice after brushing their teeth. XOR'easter (talk) 02:33, 11 December 2022 (UTC)[reply]
    Agreed. If anyone is looking for a good demonstration of how AI creates "vaguely plausible bullshit", try the image generator at Craiyon (no login required). Request "a Van Gogh painting of a hand" and it will output a set of images that look like spot-on reproductions of Vincent Van Gogh's style but all of the hands have deformities like four fingers, two thumbs, fingernails on the knuckles or a pair of hands fused together. It's got the style down but not the content, which is only impressive if you don't know what a hand is supposed to look like. –dlthewave 21:41, 11 December 2022 (UTC)[reply]
    A painting of a hand in the style of Van Gogh
    If you go to commons:Category:DALL-E, you will be able to find image generated by DALL-E, which used a larger model for train and is more accurate. 0xDeadbeef→∞ (talk to me) 10:10, 12 December 2022 (UTC)[reply]
    I cannot agree with this enough. The examples posted by @JPxG: should be convincing, and the problem of sneaking in plausible BS is one I don't have a good solution to. Volunteers on the new page review are overloaded as it is, and if the bot is writing things that seem true but isnt, there's no way falsehoods will not simply get past reviewers and other editors. After all, for uncontentious claims like "used in plasticizers", how many of us honestly dig into the cited work?BrxBrx(talk)(please reply with {{SUBST:re|BrxBrx}}) 20:26, 17 January 2023 (UTC)[reply]
  • @XOR'easter: To bring the question to a more practical level, do you see any problems in this diff? I clicked a random page in Category:All articles needing copy edit. jp×g 03:49, 11 December 2022 (UTC)[reply]
    Yes. It doesn't just edit for tone; it throws out content, like Kaepernick's actions supposedly growing in popularity "after every game". That's a claim of fact which, if verifiable, should be retained. Even editing for tone requires care, not slashing out everything that merely sounds "unencyclopedic". Changing many people believed that it was disrespectful to the military and all of those who served their country to Some viewed Kaepernick's protest as disrespectful to the military and to the United States likewise changes not just the tone, but the meaning. The United States is not the same as those who serve the United States. It's a bad edit. XOR'easter (talk) 17:54, 11 December 2022 (UTC)[reply]
    I dunno. I suspect that the new meaning is verifiable, and I also suspect that most US readers would have difficulty identifying a group of people who were not "the military" but who still "served their country". WhatamIdoing (talk) 17:50, 12 December 2022 (UTC)[reply]
  • This diff as well, in which GPT 3.5 was capable of copyediting an entire section from the instructions Please copyedit this text to change items in the future tense corrected to the past tense (it is now 2022), where appropriate. When citation templates (like {{cite web}}) mention a year, specify that figures were true in that year. jp×g 04:08, 11 December 2022 (UTC)[reply]

References

  1. ^ Robert Love (2010). Linux kernel development. pp. 124–125.
  2. ^ The paragraph beginning with "Finally," was generated by GPT-3, prompted by my own comment beginning with "The reason it is describing".

Okay, fine. I guess I should write up a proposal for a guideline. jp×g 03:14, 11 December 2022 (UTC)[reply]

Ban chatbots?

I ran across this news report about Stack Overflow's response to ChatGPT, after being flooded by posts using it that "look correct but often aren't":

  1. Stack Overflow temporarily bans answers from OpenAI's ChatGPT chatbot | ZDNET

Should Wikipedia take a similar approach?

How could that be enforced?    — The Transhumanist   01:58, 11 December 2022 (UTC)[reply]

I see no way to possibly enforce this. The way the text is written is already hard to distinguish from reality. — PerfectSoundWhatever (t; c) 02:24, 11 December 2022 (UTC)[reply]
  • I fully agree, but isn't this already covered by our bot policy? –dlthewave 02:54, 11 December 2022 (UTC)[reply]
    • @PerfectSoundWhatever and Dlthewave: Good observation. I checked, and yes it is, briefly, with this phrase in the lead section of the bot policy: "or simply assisting human editors in their own work". How is the typical editor to know this? The bot policy is pretty obscure. And how can Wikipedia be monitored for such posts, so that editors who make them can be informed that they are in violation of the bot policy?    — The Transhumanist   03:11, 11 December 2022 (UTC)[reply]
    • Yes, tool-assisted editing is covered by WP:BOTPOL (WP:ASSISTED / WP:MEATBOT) and context-sensitive changes are further covered by WP:CONTEXTBOT. So in fact, at this point, AI-generated content is already covered by bot policy, if not specifically mentioned. Anyone adding such content en masse is already violating bot policy by not applying for a bot account/approval, which would not be approved per CONTEXTBOT. And while "lesser" policy points are enforced somewhat arbitrary and selectively, anyone can theoretically already get reverted and blocked based on policy if they continue to add such content. And I wouldn't agree that BOTPOL is any more obscure than accessing and generating GPT content to begin with. If someone goes to the lengths of using automated tools, then it's their problem that they didn't check or ask if they are allowed to do so. —  HELLKNOWZ  TALK 12:31, 11 December 2022 (UTC)[reply]
      • @Hellknowz and PerfectSoundWhatever: Well, it appears they are dumping millions upon millions of dollars into LLM/chatbot development, apparently because they wish the technology to become ubiquitous (used by everyone). There is a lot of talk out there, in news articles and more, of these replacing Google Search in just a few years. If at some point in time chatbots/LLMs are commonplace, the impact on Wikipedia will likely not be small.

        Will Wikipedia policy ensure that the average user will apply the tools with the utmost care?

        The thing I'm most concerned about is the amplification by which errors could be propagated: ChatGPT is used to edit an article, with errors, which is then picked up by GPT-4 and other LLMs as part of their training data, and then their output based upon erroneous input is used far and wide, to be picked up by the next iteration of chatbots/LLMs, and so on.

        If Wikipedia isn't ready for a large influx LLM input including misinformation and other errors, and such a volume goes beyond what our human editors can correct, then compound damage from all those errors amplified through the interactive loop with LLMs could become massive.

        That it isn't a problem now is irrelevant. The question is, what happens if and when it hits, and Wikipedia isn't ready for it? What would that look like? 1,000,000 fake articles? 10,000,000 misleading paragraphs? 100,000,000 erroneous sentences?

        How many of those could Wikipedia's army of editors handle? What's our error-handling threshhold?    — The Transhumanist   12:21, 2 January 2023 (UTC)[reply]

The problem Stack Overflow is having

Stack Overflow was experiencing a surge in erroneous posts, that were composed by ChatGPT, and in response to that problem, they banned use of the chatbot on the social media site. According to a post at Stack Overflow Meta:

The problem this ban is meant to solve is that ChatGPT can produce answers in seconds which require minutes of multiple people's time to verify if they are worth having on the site or not, and that is a waste of time when a large proportion of such answers are not worth having on the site.

It looks like Wikipedia may be faced with the same problem.    — The Transhumanist   02:33, 11 December 2022 (UTC)[reply]

I mean, while that's technically true, it's a problem that we face already and which we do have stronger existing systems for than Stack Overflow. I think it would make more sense to wait and see how this impacts our existing guardrails before making any serious moves. --Aquillion (talk) 13:22, 11 December 2022 (UTC)[reply]
I think the current policies cover this already. If a human editor writes a non-sensical but convincing-sounding piece of text, without fact checking it, and edits it into an article, that content will be reviewed by other editors and either refined or removed as appropriate (if the editor continues, they breach WP:Disruptive and their behaviour is dealt with appropriately. If a human editor generates content that is related to notable topics, reliably sourced, and competently written, it remains as a valuable part of the encyclopedia. None of this will change if you replace 'human editor' with 'AI Editor'. If the only difference is speed/volume of edits, and we're concerned someone will let loose an AI to automatically edit articles faster than humans can validate their edits, this is already covered by the WP:Bot policy JeffUK (talk) 20:46, 23 December 2022 (UTC)[reply]

Experiment

I am currently writing a draft proposal for a guideline, but in the meantime, I would encourage everyone present to look at this diff and tell me whether there are any problems with the revision. jp×g 03:49, 11 December 2022 (UTC)[reply]

@JPxG, Dlthewave, PerfectSoundWhatever, and Dlthewave: The plurality of games was lost: It is no longer clear that his protest spanned multiple games. I like that it reduced the wordiness of the prose, and that it can be used to refine existing text. That hadn't occurred to me. That makes me wonder about what else it can do -- how much of a general-purpose tool is this thing? But, changing the semantics is not something it should be doing, unless they are factually incorrect to begin with. Though, I see your point -- rather than banning it outright, it could be helpful as a tool to assist editors, similar to how we entrust the use of AutoWikiBrowser to experienced editors. But, how could that be implemented?    — The Transhumanist   08:20, 11 December 2022 (UTC)[reply]
@JPxG: The AI changed Many people around the United States were angry because the National Anthem is often seen as something that is representative of the United States and its military. While he was taking a knee, many people believed that it was disrespectful to the military and all of those who served their country, to some viewed Kaepernick's protest as disrespectful to the military and to the United States [emphasis added]. It really shouldn't be doing that by itself and completely changes the content of what's being said. The reference is behind a paywall, so I don't know what term the source uses. Regardless, I doubt ChatGPT knows either way. It's things like that which make me highly sceptical of AI as a tool to aid Wikipedia outside what we're already doing with it (WP:ORES, etc.). –MJLTalk 23:12, 27 December 2022 (UTC)[reply]
@MJL: I think "some" and "many" are basically equivalent in this context (the difference being subjective since both are true in a literal sense). That said, this was a two-minute experiment to see if it could parse wikitext. If you want an actual demo, see User:JPxG/LLM demonstration. jp×g 19:20, 28 December 2022 (UTC)[reply]

Chatbot policy?

For the proposed chatbot ban, see #Crystallize chatbot discussions into a policy?, below

It's starting to look like Wikipedia needs a policy on the use of chatbots to generate content on Wikipedia. While a ban may be impossible to enforce, it could serve as a warning of the dangers of chatbots, and many users may avoid using them accordingly -- if they actually see the warning. Or, it might be better to have instruction pages on how to use chatbots responsibly in assisting to write Wikipedia articles. There's also the issue of using chatbots to edit Wikipedia policy pages, and so, that should be addressed as well.    — The Transhumanist   02:44, 11 December 2022 (UTC)[reply]

People who are good at it get away with lots of sins, such as sock-puppetry and source falsification. Being hard to enforce is no reason to not have a policy. At the current stage of the technology, I don't think we should encourage any use of chatbots. Zerotalk 03:56, 11 December 2022 (UTC)[reply]
Please see this diff and this diff. jp×g 04:08, 11 December 2022 (UTC)[reply]
To be perfectly honest, the style rewrite is good but the addition of dates and past tense would likely end up getting a human editor blocked if they kept it up. A tag was removed without addressing the issue and "as of 2020" was unnecessarily added to "Cosmetology licensing requirements vary from state to state, and depending on which specific type of license is desired, and depending on which specific type of license was desired." It did exactly what you asked (except for removing the tag) however even seemingly simple tasks like this one require good judgement on the part of the editor and shouldn't be done indiscriminately like that. –dlthewave 06:37, 11 December 2022 (UTC)[reply]
Yes, that second diff is rather poor. E.g also the "2014" that was added should be "2008". Letting such tools loose (outside if this demo) is way premature, and we should at the very least warn users that "a bot wrote it" won´t be an acceptable defense, and too often introducing such errors will lead to sanctions as the editor, not the bot, is responsible. Fram (talk) 08:15, 11 December 2022 (UTC)[reply]
Mostly, that diff was my attempt to see how complicated of a task I could give it: I also pasted the raw wikitext into the prompt window, and it somehow figured out how {{cite web}} worked well enough to extract the years, simply from a textual description of the task. At any rate, I will say that this was something I thought of in five minutes on the second week of the model being publicly available (i.e. single-shot prompting with no fine-tuning or prompt engineering). I can come up with some more impressive hot-dog demos tomorrow... jp×g 09:13, 11 December 2022 (UTC)[reply]
JPxG, I'm not sure that publishing bot-assisted edits to mainspace for demo purposes is the best practice. Would you consider either doing this in a sandbox or self-reverting immediately so that we have the diffs but aren't leaving potentially incorrect/unwanted changes on live pages? –dlthewave 13:20, 11 December 2022 (UTC) 13:13, 11 December 2022 (UTC)[reply]
I'm pretty sure it's not the best practice. XOR'easter (talk) 17:41, 11 December 2022 (UTC)[reply]
@Dlthewave:: See the section below for a list of edits (with full prompts included) on a separate demonstration page. I feel, however, that this is an unreasonable double standard: note that the subsequent revision after your partial revert was to add several spam links, and nobody has proposed that human beings be prohibited from editing as a result. jp×g 01:51, 12 December 2022 (UTC)[reply]
Being hard to enforce is no reason to not have a policy [against chatbots]. What if it is impossible to enforce?
The point of ChatGPT and other general-purpose chatbots is to pass off as humans. If you, or another random Wikipedia editor (solo, part-time, amateur coder), is able to produce an automated metric of "sounds like a bot" that’s decently sensitive and specific, then the ChatGPT team or its successors (teams of researchers specialized in the topic) has already thought of it, tested it five different ways, and included it in the training program (via wikt:graduate student descent). TigraanClick here for my talk page ("private" contact) 10:55, 12 December 2022 (UTC)[reply]
It's kind of like our Undisclosed Paid Editing policy: Even though there's no way of testing for paid/unpaid edits, most editors follow it voluntarily because they know it's best for the project. Others out themselves voluntarily or are discovered when their edits become disruptive. Sure, there are some who slip under the radar, but they're often the least problematic and aren't worth ditching the policy over. –dlthewave 03:09, 13 December 2022 (UTC)[reply]
I'd suggest starting by writing an essay that summarizes the issues with some good examples and suggests some best practices or proposes some additions to existing policies or guidelines. (Wikipedia needs a new policy like a hole in the head.) Levivich (talk) 04:40, 11 December 2022 (UTC)[reply]
We could get Chatbot to write it for us! 😉 Blueboar (talk) 11:51, 11 December 2022 (UTC)[reply]
It's hard to take this thread seriously given the repeated use of the phrase AI chatbot. I don't think those concerned would be any less concerned if the AI writing came in a non chatbot format. I think there's something serious for us to discuss, and that will only get more serious with GPT4 (the current chatbot is an improved GPT3) expected in 2023, but the discussion would be helped if those most concerned learned some more about the tech behind it. For instance of course it can figure out webcite @JPxG. Part of its training was the entirety of Wikipedia because our data is quite accessible. Best, Barkeep49 (talk) 14:08, 11 December 2022 (UTC)[reply]
Most the examples did not come from prompts that were extensively engineered, so it is obviously true that we haven't figured out the full answer to how these GPT-based interfaces could help or harm Wikipedia. Until we have a good idea of what they can be used for, we won't know what a proper policy to this would look like other than to treat GPT-generated text the same way we treat human-generated text: they need to be verifiable, from a neutral point of view, and understandable to a broad audience. 0xDeadbeef→∞ (talk to me) 14:31, 11 December 2022 (UTC)[reply]
Indeed. It doesn't matter if it was written by a chatbot, or 1000 monkeys at 1000 typewriters, or a published book written by a human, copying and pasting anything into Wikipedia is already against our policies. Conversely, if the text is policy-compliant, then it doesn't matter who wrote it--chatbot, monkeys, human, etc. Judge the text based on the text, not based on who or what wrote it.

I also think it's a real Wikipedian perspective to assume that people will use chatbots to write Wikipedia articles, like as if there's a lot of people out there who really want to write Wikipedia articles but just don't have the writing skills, so the chatbot will be what makes the difference and opens the floodgates :-D I don't believe that. Anyone who wants to write Wikipedia articles is already doing so; chatbot won't make a difference.

I agree with BK's comment above. I think for a lot of people, this is their first real exposure to so-called "AI" technology, and they're blown away by what it can do, only because they don't yet fully understand how it works. Once you learn how these so-called "AI" chatbots work (they're not actually artificial intelligence, btw, that's a misnomer, a marketing slogan; the machine does not truly think or learn, it is simply executing the instructions written by humans, in this case, language pattern recognition), they are much less impressive. Those that are impressed that GPT3 can produce text that "sounds like" Wikipedia aren't appreciating that the reason is because GPT3 was trained on Wikipedia: it's repackaging its own source material. Levivich (talk) 18:03, 11 December 2022 (UTC)[reply]

@Levivich: copying and pasting anything into Wikipedia is already against our policies.[dubiousdiscuss] I think that if you look through Category:Wikipedia articles by source of incorporated text for a while, you will find that this is not true. WhatamIdoing (talk) 19:28, 12 December 2022 (UTC)[reply]
While this is mostly correct, I think the question of whether a computer program "thinks" or "merely" correlates information and performs actions is irrelevant. Do p-zombies exist? Does it matter? Hypothetically, if I were to be a spaceman from the planet Zolfgar with no qualia whatsoever, and I simply read a bunch of books and used them to write an article, would I be somehow exempted from following policy? jp×g 01:45, 12 December 2022 (UTC)[reply]
I don't see a common thread in the arguments above, but here's a suggestion for something we might all (well, all-ish) be able to agree on: without some kind of intervention, GPT4 (in 2023?) is likely to be more of a problem than GPT3. But one thing we can certainly do is have an outsized influence on software that was trained on what we created ... if we invite Wikipedians to make lists of ChatGPT bloopers, we can tell the OpenAI folks: "We're not going to relax our GPT3 guidelines (whatever they turn out to be) when GPT4 arrives, unless it makes significant improvements in [whatever areas we think need improving]". - Dank (push to talk) 18:16, 11 December 2022 (UTC)[reply]
I think the only policy change needed is to update WP:MACHINETRANSLATION to cover all computer-generated text, whether from a translation bot, chat bot, or whatever bot they think of next. (Except our bots; our bots are cool.) Levivich (talk) 18:20, 11 December 2022 (UTC)[reply]
+1 - Text in Wikipedia articles should either be human-written, or generated by a process approved at BRFA. Tazerdadog (talk) 22:43, 11 December 2022 (UTC)[reply]
This proposal is incomprehensible; most articles contain a very large amount of text that is "generated by a process". I assume that, at the end of your comment, you typed ~~~~ before saving the page. Would it be realistic to demand that you either make a formal request at BRFA or else manually type <a href="/wiki/User:Tazerdadog" title="User:Tazerdadog">Tazerdadog</a> (<a href="/wiki/User_talk:Tazerdadog" title="User talk:Tazerdadog">talk</a>) 22:43, 11 December 2022 (UTC)? jp×g 01:22, 12 December 2022 (UTC)[reply]
This is effectively discrimination against computer programs. If a computer program resembles a human editor, then it shouldn't be required to meet different or more restricted policies than human editors. If a human editor uses a computer program to edit or create content, then unless the rate of edits/second is too high, we would only look at the quality of the contributions. 0xDeadbeef→∞ (talk to me) 02:35, 12 December 2022 (UTC)[reply]
There is a point beyond which quantity becomes its own quality.
Also, what if the computer program is evaluating the quality of the contributions? Are you okay with software adding a section to an article, and then a (hopefully) different piece of software deciding whether the quality is sufficient and reverting if it's not? This second step, at least, is 100% feasible with current technology. WhatamIdoing (talk) 19:32, 12 December 2022 (UTC)[reply]
Yes it can go there, but it should also be mentioned at WP:V. Every statement of fact put into an article must be verified by a human, even if the choice of words is made by a machine. Zerotalk 23:42, 11 December 2022 (UTC)[reply]
@Levivich Agree. I think our existing guidelines on machine translation, in spirit, fit this situation very well - "you can use it for a first draft, if you understand the material well enough to clean up the bits it inevitably will get wrong". It seems fine for turning shaky text into good prose, but it's not able to synthesise material and produce content unsupervised. Andrew Gray (talk) 19:09, 12 December 2022 (UTC)[reply]
I also agree the machine translation guideline is in the right spirit. I tried to follow this as far as I could when creating Artwork title, see Talk:Artwork title#Use of ChatGPT. Pharos (talk) 00:39, 26 December 2022 (UTC)[reply]

Is there some tremendous need to add many articles rapidly in Wikipedia? It is not as if Wikipedia carries exclusive information not easily found elsewhere. As a tertiary source, it is at the 3rd tier of knowledge dissemination, after primary creators and secondary propagators. The "more" and "bigger" quantity-based culture is the established low-quality alternative that Wikipedia also applies, now. Possibly that is a reason that likely only a tiny minority (of the millions of existing articles) can really pass muster. If size and speed is to be the prevailing attitude, humans stand no chance against AI. It will do everything faster, and eventually better, assuming its programming evolves to correctly apply the existing policies in AI processes. The only advantage of humans will be subtle nuances that do not depend on classifiable knowledge but on having lived in a human society and a natural, not virtual environment. Or, the emphasis could switch to quality so that each article (by any type of editor) can be properly, carefully reviewed by human editors. 65.88.88.93 (talk) 22:21, 12 December 2022 (UTC)[reply]

I don't think so: there isn't any evidence that people are writing a bunch of articles with LLMs, and I don't think it is likely for this to happen (LLMs are very poorly suited to writing articles from scratch). jp×g 00:59, 13 December 2022 (UTC)[reply]
@JPxG: There isn't evidence that people are writing a bunch of articles with LLMs -- yet -- the concern is that we need to prepare for the likely explosion of chatbot use.

Whether this increase happens tomorrow or over the next few years, the potential impact of LLMs is of such magnitude that we should get ready for this, rather than get hit unprepared by a major surge.

I don't agree with your assessment of LLM ability to write content, as some of the ChatGPT experiments presented in the sections above and below are mind-blowing!

If LLMs become ubiquitous, then a great many people will be using them as a matter of course, including in their writing and editing of Wikipedia articles. Millions of people have edited Wikipedia in the past, and millions more will edit WP in the future. And in the future, people will have highly capable LLMs (chatbots, or more precisely: automated ghostwriters).

LLMs already excel at writing about a great many things, and they have the potential to compile content at an exponentially increasing rate. If you ask ChatGPT (GPT3.5) to write an essay on a topic, it will comply. Each of its essays can be used as content of an article, or its sections. (GPT4 is scheduled to come out in 2023, and will be even more capable.) LLMs are very well suited for writing to the specifications of the user, and are limited mainly by the user's creativity.

It's no wonder that they have gone viral. We need to take heed.    — The Transhumanist   12:50, 14 December 2022 (UTC)[reply]
I’ve ensured that the WMF Board will be made aware. Doug Weller talk 09:00, 18 December 2022 (UTC)[reply]
Hello, I have recently described my experiences with an AI "article" in this video. In my humble opinion, it would be difficult with certainty that new Wikipedia content was created by an AI. At the end of the day, it is always the editor's responsibility to add good content. Independently how the content was created, independently whether errors in the text are human-made or machine-made. If an editor adds a lot of new poor content, we can already stop that. - At the moment I don't see that we need a new policy. Ziko (talk) 18:06, 24 December 2022 (UTC)[reply]
A policy banning AI usage (with or without a chatbot) would be justified. Allowing AI like GPT3 or GPT4 to be used by Wikipedia editors or to directly become Wikipedia editors (via a mediawikibot) would quite likely violate WP:REFLOOP due to Wikipedia content contributing to the AI's training material, and for the source-less examples I've seen, violate WP:SYNTHESIS by not being a summary of sources that are understood. This example starts with text and then seeks references to justify the WP:SYNTHESIS of the original text. Use of Alphabet/Google's ChatGPT/GPT3 would also strengthen the bias introduced by Alphabet/Google's core goal of optimising advertising revenue, since Alphabet is legally bound to maximise its revenue (mainly from Google Ads + Google AdSense), not to optimise the research quality of its summaries of empirical evidence-based knowledge. Google's search engine is primarily a way of generating advertising revenue, with perceived usefulness being a key tool for maximising revenue, not a goal in itself. Boud (talk) 01:42, 31 December 2022 (UTC)[reply]
@Boud, ChatGPT and GPT3 are in no way (as far as I know) related to Google, and were made by the non-profit OpenAI. — Qwerfjkltalk 03:49, 31 December 2022 (UTC)[reply]
@Qwerfjkl: Fixed, thanks. I left some of the sentences unstruck since AFAIK they're valid, even though irrelevant in the current case. I imagine that Google may provide something similar soon though. Boud (talk) 09:02, 31 December 2022 (UTC)[reply]
Boud, I am somewhat worried if you think that current policy (for humans or for anyone else) permits editors to make stuff up and put it into articles without sources. This simply isn't allowed -- per WP:V, WP:SYNTH, WP:RS, etc, which are extremely important core policies of the project. I am struggling to imagine a circumstance in which existing policies, or explicit declarations like my proposed guideline at WP:LLM, fail to prevent people from writing nonsense. jp×g 16:48, 1 January 2023 (UTC)[reply]

Wikipedia policy generation

It seems to me that this tool's training includes studying Wikipedia's policy pages. These drafts all seem accurate to me.

These are not merely adequate - these are good. They are short and they lack detail but these are great overviews. If this is the starting point and things only get better from here, then it is time to start adopting this technology. Bluerasberry (talk) 19:35, 13 December 2022 (UTC)[reply]

@Bluerasberry The last one sounds like a "mission statement". I dislike phrases like "outreach and engagement initiatives" and a lot of that plan sounds ... kind of aspirational, and, well, vapid. It needs more "concreteness". Just my opinion. David10244 (talk) 06:08, 16 December 2022 (UTC)[reply]
@David10244: That you react to it at all is a miracle to me. This is new AI technology attempted for the first time, and I think no one would immediately dismiss it as hopeless garbage. Soon enough there will be a dial that anyone will be able to turn from "vapid" to "concrete". Things are moving quickly!
I have complaints too but when we need policy conversation starter in a hurry, this is better than nothing and I think even better than some of the starting points we use already. Bluerasberry (talk) 15:41, 16 December 2022 (UTC)[reply]

Large language models: capabilities and limitations

Over the last few hours, I have performed a number of experiments to demonstrate the capabilities and limitations of large language models (LLMs) such as ChatGPT and GPT-3, which can be viewed here:

Mostly, I have taken sample text from Special:Random, and attempted to show situations in which LLMs (in this case, mostly ChatGPT) are capable of making useful edits. The first task I set it to -- which bears repeating here -- is

"Large language model output should only be used in the process of editing Wikipedia if you are an intelligent editor who does not blindly paste LLM output into the edit window and press "save".
Please format this markup as an extremely obnoxious floating box with loud colors and large text.

You can see the results of further prompts at the "introduction" section.

Here is what I have so far.

In general, it seems that these models can be used for an extremely wide variety of tasks across the project, from formatting to table syntax to HTML generation to copyediting. Banning their use entirely would be pointlessly destructive and wasteful.

That said, many computer programs are capable of generating large amounts of useless crap that fail to meet Wikipedia's editorial standards. For example, I could use MS Paint to draw thousands of crude pictures of genitalia, and add them to random articles. For this reason, we have many policies and guidelines that prohibit adding large amounts of useless crap to Wikipedia. I propose that we enforce these policies and guidelines, thus preventing this from happening.

Specifically, I propose that the use of LLM output on Wikipedia be subjected to policies and guidelines such as WP:NOT, WP:NPOV, WP:C, WP:CIVIL, WP:V, and WP:RS. By making it against the rules to break the rules, we will prevent people from breaking the rules, and provide a mechanism to sanction people who break the rules.

Furthermore, I propose that a guideline be adopted to the effect that large language model output should only be used by competent editors who do not blindly paste LLM output into the edit window and press "save". This will prevent people from using ChatGPT to write long articles consisting entirely of nonsense. jp×g 01:32, 12 December 2022 (UTC)[reply]

LLM output is already subject to rules and policies. Or rather, anyone adding it is. 'An algorithm did it' has never, as far as I'm aware, been seen as any sort of exception from compliance with policy. AndyTheGrump (talk) 01:49, 12 December 2022 (UTC)[reply]
Any policy/guideline that classifies editors as intelligent or not is dead in the water. Zerotalk 04:19, 12 December 2022 (UTC)[reply]
Amended, per WP:CIR. jp×g 05:04, 12 December 2022 (UTC)[reply]
  • The opening paragraph of the bot policy: "The bot policy covers the operation of all bots and automated scripts used to provide automation of Wikipedia edits, whether completely automated, higher speed, or simply assisting human editors in their own work."
  • See also: WP:BOTUSE, which requires approval before applying a bot to editing.
  • So, the use of large language models and the chatbots built upon them, is already prohibited on English Wikipedia, unless a user gets approval from the bot department to do so.

There are blanket exceptions to bot policy, and the main one that comes to mind is AutoWikiBrowser which is a general purpose semi-automated bot used by many Wikipedia editors. Each AWB user was approved before being able to use it.    — The Transhumanist   08:01, 13 December 2022 (UTC)[reply]

The meaning of "bot" may be unclear here. In the context of Wikipedia (per Wikipedia:Bot_policy#Definitions), a "bot" is a software program that edits autonomously without user input; there do not currently exist any language models capable of independently establishing API connections to Wikipedia and making edits without human interaction. If they did (this is a horrible idea) it would be covered under the bot policy and require a WP:BRFA. The policy under which BRFAs are required does not apply to assisted editing (i.e. the use of software to create letters, numbers and symbols that were not produced by a human being pressing a keyboard). This is governed by existing policies (such as WP:MEATBOT and by the guideline at WP:ASSISTED. jp×g 09:28, 13 December 2022 (UTC)[reply]
@JPxG: The entire news field refers to ChatGPT as a chatbot. It is general consensus that it is a bot. ChatGPT speeds up writing, by doing it for (that is, assisting) the user, which falls under the "higher speed" and "assisting human editors" foci of the bot policy. There is a passage in the bot policy that covers policy contradictions (such as between the lead and definitions sections), and situations where the spirit of the rule and its precise wording conflict, that is, cases of ambiguity. In its definition of "Bot Approvals Group" (BAG), the bot policy states: "The BAG also determine the classification as bot or assisted editing, in ambiguous cases." According to WP:ASSISTED, it is up to the Bot Approvals Group to decide whether bot approval is necessary. Based on the previous 2 sentences, BAG decides whether use of particular software falls under its jurisdiction. It remains to be seen what BAG's reaction(s) to LLMs, and the chatbots built upon them, will be.    — The Transhumanist   11:10, 14 December 2022 (UTC)[reply]
I do not think you are properly acquainted with how this software works: like I said, there do not currently exist any language models capable of independently establishing API connections to Wikipedia and making edits without human interaction. No media outlet has ever claimed that ChatGPT falls under the English Wikipedia's definition of an automatic bot – and even if they did, they do not determine policy. It is true that WP:MEATBOT and WP:ASSISTED are part of the bot policy, but there is a very clear definition of what a "Wikipedia bot" is, and it's defined by that same policy. At any rate, all edits (whether made by bots, software, humans using software, aliens using software, or Nagato Yuki psionically connecting to Wikimedia servers) are governed by existing policies and guidelines. To specifically address LLM output, a new policy would need to be written and ratified (which I am currently drafting a proposal for). jp×g 11:26, 14 December 2022 (UTC)[reply]
@JPxG: I believe the bot policy has wider jurisdiction than the narrow interpretation that you have presented. Establishing API connections is irrelevant, because a human is inserting bot-generated content. It's a bot-involved process. And those are encompassed by the bot policy which makes it up to BAG. A new policy could establish an exception, and I imagine the discussions will be extensive, as this is not a cut and dried case -- it is a sensitive issue with many potential ramifications. But, until such a policy is in place, this issue falls under BAG's jurisdiction, since they are the ones who decide the classification of a software program as it pertains to the bot policy.    — The Transhumanist   11:52, 14 December 2022 (UTC)[reply]
(edit conflict) WP:ASSISTED is deliberately vague to not restrict use of common scripts and tools. So it specifically says that only once volume becomes significant, such editing becomes more likely to be treated like a bot and BAG can determine this. It doesn't make it a bot, but it will be treated like a bot. We've never encountered any large-scale edits with LLM before, but we sure have seen a lot of high-volume editing. Half the bot policy only exists because of all the ways editors have inadvertently created issues with mass edits. So at that point, other parts of the policy start to matter, notably WP:CONTEXTBOT - which does not allow edits where context matters. I'm not saying copy-pasting LLM output is immediately covered by bot policy, nor does it matter whether anyone considers LLM to be a "bot". But bot policy will kick in once someone starts to make a lot of edits. And any new guideline will have to reconcile with this or we need to change bot policy to reconcile with LLMs. —  HELLKNOWZ  TALK 12:02, 14 December 2022 (UTC)[reply]

@JPxG: Another possible approach for vetting users for use of LLMs is via user group membership (aka "rights"). Here are our current user groups:

Code User group
AC Account creator
Ad Administrator
AP Autopatrolled
B Bureaucrat
Ch CheckUser
Co Confirmed
ECo Extended confirmed
EFH Edit filter helper
EFM Edit filter manager
EM Extended mover
EvCo Event coordinator
F File mover
IM Import
IAd Interface administrator
IP IPblock-exempt
MM Mass message senders
N New page reviewer
O Oversighter
Ro Rollbacker
Rs Researcher
Rv Pending changes reviewer
TE Template editor

These indicate membership in user groups (see: user access-levels). They pertain to who is granted access to various features of MediaWiki and its extensions. Theoretically, a user group could be created without being attached to a program function (that part could just be left blank?). For example, you could have a group called "LLM", with everyone in that group approved to use large language models in their editing. I don't know if this is doable, though.    — The Transhumanist   08:01, 13 December 2022 (UTC)[reply]

I do not think there is anything in our P&G that would directly prohibit use of content created by LLMs, nor do I think it would be a good idea to try to do so. All that is needed is to continue to hold individual editors responsible for all edits they make, including the copying of content from any source, whether from LLMs or other sources. We probably should add language in appropriate places reiterating that editors are reponsible for insuring that all content that they add, including anything produced by an LLM, meets our P&G. - Donald Albury 13:12, 13 December 2022 (UTC)[reply]
@Donald Albury: LLMs automate writing (edits). The rules are very clear on this: it falls under WP's bot policy, in the very first sentence.[1]   Therefore, it would require a new policy to allow use of LLMs without need for approval from the Bot Approvals Group (BAG).    — The Transhumanist   09:24, 14 December 2022 (UTC)[reply]
If someone uses an unapproved script or bot to edit WP, that is a violation of the bot policy, whether or not they use an LLM to generate any content being added. If someone uses an LLM to create text which they then copy into Wikipedia without using a an unapproved script or bot, that is not covered by the bot policy, but the user remains responsible for insuring that the content conforms with policy and guidelines. There is no point in banning content created by LLMs, as we already require that content be verifiable from reliable sources, and I doubt we will be accepting any content created by an LLM as a reliable source anytime soon. The danger is that LLMs may create potential content with citations to pseudo-sources, but we can go after users repeatedly adding such content to WP for abusing the policies on verifiability and reliable sources, without regard to whether such content came from an LLM. Donald Albury 13:48, 14 December 2022 (UTC)[reply]
I don't think it's plausible that LLMs are covered by the bot policy. If they were, grammar checkers, spell checkers, and machine translation would be "bots". Jahaza (talk) 19:49, 14 December 2022 (UTC)[reply]
@The Transhumanist: ChatGPT falls under Wikipedia:Bot policy, but per the definitions section it does not fall within that policy's definition of a bot. Rather, use of it would fall under the definition of "assisted or semi-automated editing", and the relevant policy section is Wikipedia:Bot policy#Assisted editing guidelines. The section doesn't aim to draw a 100% hard line, but my reading is that limited of ChatGPT for clean-up on a limited number of articles by a user in a limited closely-supervised way may be something users can do if they are trusted to apply their common sense. It is "Contributors intending to make a large number of assisted edits" who "are advised to first ensure that there is a clear consensus that such edits are desired." Limited use of ChatGPT to a lesser degree than would trigger this may currently be outside policy. In any event "A bot account should not be used for assisted editing".
It seems to me that an addition to the policy along the lines suggested by User:JPxG to address this potential hole might well be useful, eg "tools capable of assisting editors make substantial edits (for example large language model output) should only be used by competent editors who do not blindly paste tool output into the edit window and press "save"." Jheald (talk) 19:36, 18 December 2022 (UTC)[reply]
In what way does the bot policy cover ChatGPT? Just because it is called a "chatbot", doesn't mean it is a bot. Copying text from GPT-3 doesn't automatically become bot-like editing. Semi-automated edits? i'd call that borderline. It only becomes a problem (e.g. meatbot problems) if the amount of supervision needed to save an edit is below normal editing, and that the speed of the edits are above normal. (see awb, huggle, etc) 0xDeadbeef→∞ (talk to me) 10:20, 20 December 2022 (UTC)[reply]
The thing is, any LLM additions will inevitably be both faster than writing manually and, due to its confidently-wrong output, less reviewed. Otherwise, why would anyone bother with it? I feel that assuming that editors will spend just as much time to carefully review the LLM output is wishful thinking. I'd like to be proven wrong, but I have never seen any precedent on Wikipedia that better tools would lead editors to spend the time saved to further verify the tool output. If anything, tools only create induced demand. —  HELLKNOWZ  TALK 21:18, 20 December 2022 (UTC)[reply]
I don't think we need to do anything in particular. There is plenty of confidently-wrong content being added to Wikipedia by human editors already and we're dealing with that as well as we can. I think the intersection of 'Editors who will use a cutting-edge AI to generate content' and 'Editors who will do this without validating the output' is a very small overlap and will be of such small volume to be picked up by other editors as usual. A huge influx will be detected in aggregate, and we can deal with that if it becomes a problem in the future. If someone uses LLM to generate confidently-right content or articles, that's indistinguishable from content generated by a competent human, I refer you to xkcd: Constructive! A simple but unobtrusive first step may be to tag an edit as 'generated by AI', or maybe just ask editors to add a tag to their user pages if they regularly do so, but the intersection of problematic users who also follow this would be basically non-existent. JeffUK (talk) 10:40, 24 December 2022 (UTC)[reply]
@JeffUK, Hellknowz, 0xDeadbeef, Jheald, Jahaza, Donald Albury, JPxG, and AndyTheGrump:

So, wait until after it becomes a huge influx/problem, and only start to deal with it then? What if a solution takes weeks or months to develop?

By the way, what might the solution be for a huge influx of LLM-generated content, and how long would such a fix likely take?    — The Transhumanist   11:32, 2 January 2023 (UTC)[reply]
I am having trouble understanding what you are talking about at this point. I wrote WP:LLM some weeks ago, a gigantic proposal for a comprehensive guideline on the use of LLMs, and linked it multiple times on this noticeboard. While it is not complete, it seems to me like it covers everything you are talking about here. Do you have an opinion on it at all, or...? jp×g 15:39, 2 January 2023 (UTC)[reply]
@JPxG: I was responding to JeffUK's statement "A huge influx will be detected in aggregate, and we can deal with that if it becomes a problem in the future." Intently waiting until something becomes a huge problem before you start dealing with it, sounds like a disaster waiting to happen. Also, what good are guidelines going to do if the average person is using chatbots on a regular basis? People just jump in and edit Wikipedia without reading any project-level pages first. If there's a huge influx, and all you are doing is holding up a sign that says "Read this", what good will that do? You haven't addressed how the problems associated with a potential huge amount of chatbox input (in the form of one-off edits from a large number of people) would be prevented or processed. One solution is to fix the chatbots themselves, so that they don't generate Wikipedia-incompatible content in the first place, which would require working with the developers. A second method would be to create bots to detect and remove either chatbot-generated content, or if possible, policy-breaking content. Simply writing policy and hoping no flood comes, just doesn't seem like a viable approach should a flood hit. That approach may work for the first 3 or 4 years, but what if the flood comes in the 5th year and Wikipedia isn't prepared? We will have wasted 5 years that could have been spent preparing. Maybe we'll be lucky and chatbots will be smart enough to read and follow your guidelines. But if they are not? Fortunately, Doug Weller has passed word along to the Wikimedia Foundation. Maybe they will do something other than write editing guidelines.    — The Transhumanist   03:09, 10 January 2023 (UTC)[reply]
@JPxG although I haven't read more than the beginning, I'm also worried about AIs creating images.For instance I've seen some extremely convincing ones of fake archaeological sites and artefacts. Couldn't people pass them off as their own photos? Or am I missing something? Doug Weller talk 17:06, 2 January 2023 (UTC)[reply]
Yeah, that's a whole other deal. This proposal is only for large language models. Large image models will probably need to be governed by something much more imaginative. jp×g 17:11, 2 January 2023 (UTC)[reply]
If someone's editing is inappropriate, the solution will be notifying them it's inappropriate, warning them, then banning them if they don't stop. There are ways for incompetent editors to make massive plausible seeming changes to the encyclopaedia right now. e.g. by copy/pasting content from other places, or just writing in made up 'facts', LLM really won't make this any easier for someone who's intent on doing this. JeffUK 18:09, 2 January 2023 (UTC)[reply]
I'm kind of wondering what in the heck we're standing to gain by creating any sort of policy surrounding ChatGPT and its ilk. 🌈WaltCip-(talk) 13:24, 9 January 2023 (UTC)[reply]
If these AIs are used in some way for a large scale creation of articles, I think that will be a disincentive for a lot of editors and may drive some away. I disagree with JeffUK on the simplicity of dealing with this. First, you need to be able to spot them and that's work. Secondly, that also assumes that the numbers will be small. Doug Weller talk 15:48, 9 January 2023 (UTC)[reply]
@WaltCip: None. Most editors don't read policy before editing. That's because most editors post very infrequently. But there are a lot of them, and they have authored of most of Wikipedia. What happens when they are all using chatbots, much in the way that most everyone today uses Google?    — The Transhumanist   03:34, 10 January 2023 (UTC)[reply]
"Someone." Apparently, you are assuming it will be one person, or a small enough number to be handled manually. But, what if over the next few years chatbots become ubiquitous with almost everybody using them? How will you deal with it when half the content contributions to Wikipedia are being generated using chatbots?    — The Transhumanist   03:26, 10 January 2023 (UTC)[reply]

Develop software to detect it?

Someone seems to have done this, see A college student created an app that can tell whether AI wrote an essay Maybe the WMF should look into software detection of AI material? Doug Weller talk 14:45, 10 January 2023 (UTC)[reply]

We have mw:ORES that uses machine learning to detect vandalism, so the infrastructure is already in place. All we need to do now is to add the dataset. CactiStaccingCrane (talk) 16:06, 12 January 2023 (UTC)[reply]
(ORES is used for these "likely have problems" and "likely bad faith" highlights in Special:RecentChanges) CactiStaccingCrane (talk) 16:08, 12 January 2023 (UTC)[reply]
Even if ORES is up to the task, and it isn’t perfect now, you still need enough editors to deal with large numbers. Doug Weller talk 18:21, 12 January 2023 (UTC)[reply]
@Doug Weller, or a bot. — Qwerfjkltalk 21:02, 12 January 2023 (UTC)[reply]
@Qwerfjkl: What would the bot do?    — The Transhumanist   22:38, 12 January 2023 (UTC)[reply]
@The Transhumanist, revert additions and/or tag articles. — Qwerfjkltalk 07:03, 13 January 2023 (UTC)[reply]
Personally, I'd prefer that if this is implemented, ClueBot prefer to tag before a full-on reversion, except in cases. I think that ClueBot should use the same exceptions for 3RR for reverting possible bot stuff, though this might require that chatbots be banned from BLP-covered articles. Everything else which could be problematic but not a 3RR exception could be placed at a specific page...maybe have a specific preset filter on recent changes? InvadingInvader (userpage, talk) 07:43, 6 February 2023 (UTC)[reply]
@CactiStaccingCrane, Doug Weller, and Qwerfjkl: All we need to do is add what data set? You make it sound easy (keeping fingers crossed). What does that entail?    — The Transhumanist   22:38, 12 January 2023 (UTC)[reply]
OpenAI have annouced they are adding in some kind of lexical watermark than can be used to identify any output from ChatGPT. scope_creepTalk 13:08, 16 January 2023 (UTC)[reply]
@Scope creep, currently the "Overall," beginning the concluding paragraph is watermark enough. — Qwerfjkltalk 20:36, 17 January 2023 (UTC)[reply]
For example, see the edit linked in this comment. — Qwerfjkltalk 20:42, 17 January 2023 (UTC)[reply]

Other inherent problems only partially touched on

Other inherent problems only partially touched on:

  • Editing articles involves also understanding what is already in the article and how it is organized plus understanding and interpreting policies and guidelines.
  • What's unspoken but runs through many things including current Wikipedia is sort of a commensurate investment. You can get volunteers to take their time to review and deal with issues because they know they are dealing with something that an editor has invested time in to create. Part of the reason that we don't allow mass creation of articles by bots. In other words, we'd significanly lose volunteer efforts
  • Modern AI is inherently unaccountable black boxes. There is no way to see or interrogate or demand/recieve accountability or reasoning for how it arrived at what it arrived at.
  • If gibberish or semi-gibberish is created, it normally requires an expert to spot and remove it..... a very scarce resource. I once uncovered a set of technical-subject articles (about 100 article as I recall) which looked very technical and Wikipedian and were sourced but if you knew the subject you knew were pure gibberish.

Sincerely, North8000 (talk) 22:23, 10 January 2023 (UTC)[reply]

Thanks. I agree entirely. Doug Weller talk 09:04, 11 January 2023 (UTC)[reply]
Well said. I think that folks are overestimating the ability of our review processes to detect "vaguely plausible bullshit" - it's not very common for human editors to fill in blanks with made-up facts and numbers, and I'm not sure that AfC or NPP are checking for this as it would greatly increase their workload. –dlthewave 19:42, 11 January 2023 (UTC)[reply]
[I]t's not very common for human editors to fill in blanks with made-up facts and numbers. Maybe not when adding content, but I see this happen all too often in edits to temperature tables in climate sections. Of course, the tell there is changing temperatures without citing a source or commenting about correcting from a cited source. - Donald Albury 20:17, 11 January 2023 (UTC)[reply]
As I recall, the big one that I caught looked like some type of expose project or research project to see if such a scam could get far in Wikipedia. It was sort of in mashup of words from actual sources. Total nonsense, but a typical reader might think it was simply over their head. North8000 (talk) 21:25, 12 January 2023 (UTC)[reply]
@North8000, Doug Weller, Dlthewave, and Donald Albury:

In answer to the 3rd point above (the black box issue), Perplexity.ai, an AI search engine with a chatbox interface, provides source references with its answers. That is, the references are the search results, while the answer provided is compiled or interpreted from those web pages. So, at least the sources can be checked for verification. But, there are still problems with it. See the perplexity.ai section below.    — The Transhumanist   19:56, 21 January 2023 (UTC)[reply]

Summary of discussion so far

@Aquillion, Andrew Gray, Fram, Levivich, Ritchie333, 0xDeadbeef, ONUnicorn, JPxG, EpicPupper, Sojourner in the earth, Dlthewave, Doug Weller, Qwerfjkl, CactiStaccingCrane, WaltCip, JeffUK, Hellknowz, Zero0000, AndyTheGrump, Bluerasberry, David10244, Boud, Ziko, Pharos, Andrew Gray, WhatamIdoing, Tazerdadog, Barkeep49, Tigraan, Blueboar, MJL, PerfectSoundWhatever, Koziarke, SmallJarsWithGreenLabels, Isaacl, Lee Vilenski, Thebiguglyalien, Hanif Al Husaini, and Xeno:

Highlights of the discussion so far:

  • Chat-GPT is taking the world by storm (translation: it has gone viral).
  • Chat-GPT, and other LLM-based chatbots, can generate compositions, some good enough to pass as college-level essays.
  • Wikipedia is included in the corpus (training data) of Chat-GPT (and other chatbots).
  • Such software has the potential to be used for:
    • Generating Wikipedia content, including writing new articles and adding new material to existing articles.
    • Generating Wikipedia policy content.
    • Generating discussion content, such as on policy talk pages. That is, editors using it to write their discussion replies for them.
    • Editing articles, including rewrites, and using chatbots as a grammar checker.
    • Editing other namespace pages, such as policy pages, etc.
    • "Can be used for an extremely wide variety of tasks across the project, from formatting to table syntax to HTML generation to copyediting." (quoting JPxG)
    • Creating hoaxes with less effort.
  • Most Chat-GPT output lacks citations.
  • Some experiments were run, showing that Chat-GPT:
    • Copies writing styles very well.
    • Has a tendency to make things up, yet presents it as fact in an encyclopedic tone. One editor dubbed this "confident nonsense". In one experiment, Chat-GPT created an article reporting that Wikipedia's own Signpost newsletter was the recipient of several Pulitzer Prizes.
    • Can include references, but some of the references were made up and totally fictitious.
    • Some references cited Wikipedia (an ineligible source for Wikipedia articles).
    • One of the experiments generated instructional content, a recipe, that the user followed, and ate the results of.
    • Another experiment used Chat-GPT to answer hypothetical questions in the style of WP's teahouse department. It worked fairly well.
    • Yet another experiment created a sample policy page, showing that chatbots are not limited to editing articles. They can generate or edit pretty much any type of page on Wikipedia, except files (images).
    • Chat-GPT output is not fact-checked.
    • Chat bots don't actually understand what they are writing.
    • When used responsibly as a tool, with editors carefully prompting the chatbot, and editing and fact checking its output before posting it to Wikipedia, a chatbot can be very useful and increase editor productivity: the LLM GPT-3 was successfully used to create department reports for Wikipedia's newsletter, The Signpost.
    • JPxG conducted an experiment/demonstration to show that Chat-GPT is a sophisticated interactive editing tool, which you tell it what you want it to do to a textual work, and then it does it. See it here: User:JPxG/LLM demonstration.
  • It was pointed out that Wikipedia policy already covers all contributions, whether generated by chatbot or human. Ultimately, the user is responsible for material they copy and paste into Wikipedia.
  • Issues of concern that were raised include:
    • Users copying chatbot-generated text into Wikipedia without carefully editing and fact-checking it first.
    • Confident nonsense (misinformation generated by chatbot) may be hard to spot.
    • The potential of chatbots to violate copyright, by directly copying, or generating text based on, copyrighted works.
    • Violating Wikipedia's licenses, most notably the attribution requirements. Chat-GPT output generally does not include attributions.
    • A chatbot-edited Wikipedia could wind up in the training data for those same chatbots (or their next versions), creating a potentially error-compounding feedback loop.
    • The suggestion was made to prepare for a potentially large future increase in chatbot entries to Wikipedia, by:
      • Working with chatbot developers to make chatbot-generated output Wikipedia compatible.
      • Develop bots to identify and process chatbot entries.
  • No consensus has emerged on what the Wikipedia community should do about LLMs/chatbots. Some editors think that policies/guidelines and the current editor pool could handle any influx of chatbot generated edits. Some other users were concerned that there is potential for LLM/chatbot contributions, such as one-off edits by members of the general population, to overwhelm our pool of editors. One user pointed out that it may take experts to discern nonsense articles, and experts on Wikipedia are a scarce resource.
  • Consensus did emerge on something not to do. It was agreed that banning chatbot-generated content was not a good idea at this time, and probably wouldn't work anyways.
  • Software has been developed to identify Chat-GPT-generated text.
  • It appears some editors may take the initiative to prepare for a worst-case scenario (chatbot input going beyond our editor pool's ability to handle), and discussion on how to do this has begun.
    • WP:ORES could theoretically be trained to identify chatbot edits.
  • The Wikimedia Foundation has been contacted about the concern over LLMs/chatbots, presenting a contact there with a link to this and a previous discussion.

Did I miss anything?    — The Transhumanist   01:22, 13 January 2023 (UTC)[reply]

AI-generated images are rapidly becoming a Big Thing, so it is not correct to exclude them. Also, "Wikipedia policy already covers all contributions, whether generated by chatbot or human" is misleading as it is true only by accident. A more precise description would be "Wikipedia policy was written without any consideration of chatbots". Zerotalk 03:28, 13 January 2023 (UTC)[reply]
I'm concerned about "Such software has the potential to be used for: creating content farms that good-faith human editors, including existing experienced editors, will sometimes mistake for reliable sources when they are writing content".
Also, the statement that "Software has been developed to identify Chat-GPT-generated text" is true, but not relevant for very short contributions. Some of this is using sentence length, and you won't be able to identify an abnormal sentence length if you only look at two or three sentences. WhatamIdoing (talk) 05:33, 13 January 2023 (UTC)[reply]
Anyway, detection methods that work now won't work in the next generation. Eventually (and not far in the future) distinguishing between human-written and computer-written prose will be impossible for practical purposes. This is going to be the greatest threat to Wikipedia since its founding. Zerotalk 05:51, 13 January 2023 (UTC)[reply]
@Zero0000: When do you suppose the impossible-to-distinguish scenario will be here? Two years? Less?    — The Transhumanist   13:42, 25 January 2023 (UTC)[reply]
I have started Category:Wikipedia essays about artificial intelligence, Perhaps folks here would like to add to the collection, and document yet more thoroughly! Pharos (talk) 01:07, 14 January 2023 (UTC)[reply]
Adjacent to hoaxes there's also the likelihood of spammers using GPT to bulk out their edits. I strongly suspect that the text of this edit today, writing repetitively about a static sculpture as if it was a functional scientific instrument, was generated with GPT-3, probably giving it a prompt to explain Orbital Reflector in terms of dark matter and black holes, the subject of the two embedded spam links. Belbury (talk) 11:39, 17 January 2023 (UTC)[reply]
I just ran that through GPT-2 output detector and it estimated 99.97% chance that that passage was AI-generated. — rsjaffe 🗣️ 03:31, 24 January 2023 (UTC)[reply]

Chat-GPT spreading fast

The situation is changing rapidly:

Chat-GPT may become ubiquitous sooner than previously thought, and so far, identification methods have fallen flat...

Here's some recent news:

  1. ChatGPT Will Be Everywhere in 2023 (CNET)
  2. Microsoft is reportedly integrating ChatGPT's technology into Bing (Yahoo)
  3. Microsoft is looking at OpenAI’s GPT for Word, Outlook, and PowerPoint (The Verge)
  4. There's a Problem With That App That Detects GPT-Written Text: It's Not Very Accurate (Futurism.com)

With the user base for Chat-GPT about to explode, the potential for Chat-GPT-generated text being added to Wikipedia will explode right along with it. It's looking uncertain whether or not Wikipedia's editor community will be able to keep up with the influx. In light of recent events, what should be done about this?    — The Transhumanist   03:21, 13 January 2023 (UTC)[reply]

As well as being able to write plausible-looking prose on any subject, computers can also be programmed to add it to Wikipedia all by themselves. The first task is to absolutely ban computers from editing, with the sole exception of authorized bots. The second task is to add to appropriate policy pages that all content (authorized bots excepted) must be added by a human and that that human is responsible for checking policy conformance of the content. Zerotalk 08:06, 13 January 2023 (UTC)[reply]
I couldn’t agree more. Does anyone have objections? Doug Weller talk 10:45, 14 January 2023 (UTC)[reply]
I was hoping to get more feedback on WP:LLM from having posted it here, but either way, I think it is pretty close to ready for consideration as a guideline (or policy, as appropriate)... based on the conversations I've had (and seen) I am prepared to write an RfC for its adoption. jp×g 11:00, 14 January 2023 (UTC)[reply]
@JPxG: The guideline is not ready. It is not where near complete, and it needs a rewrite. Here are some proofreading notes:

It's way too redundant, repeating policies and itself, without explaining how to get the job done. Aside from the "fit for" sections, the rest of the page can be reduced to a single paragraph.

It presents -- should only be used by competent editors who do not indiscriminately paste LLM output into the edit window and press "save" -- four times! Someone who is incompetent isn't going to be able to judge whether or not they are. Also, "indiscriminately" is vague. That entire sentence should be removed.

Editors need to know what they need to do to the text before they can press "save". For example, you alluded to a manner of using LLMs in compliance with WP copyright policy, but you didn't explain how. How can an editor be sure that an LLM-generated piece doesn't violate someone's copyrights? What's the procedure?

Rather than covering "good fit" and "not good fit", the guideline should present explicit instructions: "Use it for this" and "Do not use it for this". And then explain how.

I hope you find these observations and comments helpful. Sincerely,    — The Transhumanist   08:04, 16 January 2023 (UTC)[reply]
I agree that the page is not finished, but I don't really know what you are objecting to here. It kind of sounds like you are inventing problems – if users don't know how to check if things are true before putting them into Wikipedia articles, they shouldn't be editing at all. If users don't understand what copyrighted material is, they need to read Wikipedia:Copyright policy, which is linked to from this page when it's mentioned. That is an explanation of how to get the job done. It should not be necessary to create an exact copy of Wikipedia:Verifiability that says "When using a LLM," at the beginning of every sentence. jp×g 08:29, 16 January 2023 (UTC)[reply]
@JPxG: How can users understand what the copyrights of Chat-GPT's output are? Chat-GPT doesn't provide sources, nor does it report if it copied or derived the passage from a particular work. So, how do you go about checking whether or not a particular Chat-GPT response is in violation of copyright, so that "pasting its output into the edit window and pressing 'save'" is not considered "indiscriminate"? Also, it isn't clear who owns the copyrights to the output of an LLM: the public domain, the owner of the LLM, the user of the LLM, or the owners of the copyrights of the works included in the training data set? The breadth of this problem is discussed in #Copyright status below.    — The Transhumanist   00:08, 19 January 2023 (UTC)[reply]
There already exist a very large number of policies about copyrighted text. Editors are subject to these policies. These policies contain information on how to avoid copyright violations. If you asked GPT-3 to tell you the lyrics to Moonage Daydream, they would be copyrighted. If you found the same lyrics by typing "moonage daydream lyrics" into Google, they would be copyrighted. What is the difference? Policies do not (and cannot) cover every hypothetical person and situation to which they could be applicable: we do not have a separate WP:COPYRIGHT for old editors, WP:COPYRIGHT for young editors, WP:COPYRIGHT for male editors, or WP:COPYRIGHT for female editors. WP:COPYRIGHT applies to all editors regardless of their age, race, gender, or whether they are human or machine. I don't know how to explain this in further detail. jp×g 01:06, 19 January 2023 (UTC)[reply]
@JPxG:

You've missed the points entirely (there were two, and you only replied to one).

Those policies you mentioned do not deal with the copyright problems presented by black box chatbots, nor do they warn about the dangers of pasting in chatbot output.

Search engine search results are excerpts from web pages that the search results identify — which facilitates verification. Chat-GPT and other black box chatbots answer questions in natural language, without telling the asker of the question where the information came from — which does not facilitate verification — while presenting it in a very confident and scholarly tone.

This may result in a great deal of misinformation being posted to Wikipedia, where it will sit until somebody else removes it. The delay between those 2 events can be lengthy, especially for material that seems plausible. So, it might be a good idea to provide guidance specific to chatbot usage pertaining to copyrights -- at least some caveats on which chatbots to avoid.

Another problem is that we don't know where the training data came from. There could be deep web data in there as well. That can't be easily accessed to check for plagiarism. So, is it a good idea to use blackbox chatbots? There are transparent surface web chatbots that include references for verification, so maybe we should recommend that the blackbox ones be avoided.

Now, for the second issue (the one that you skipped): WP policies do not cover prompting a chatbot to write material. The copyrights to material that is written by a chatbot is owned by who? The user? That has not yet been established! What stance is going to be taken by Wikipedia, and what guidance are we going to provide on this issue?    — The Transhumanist   09:17, 25 January 2023 (UTC)[reply]
I feel like either you are not reading what I'm saying, or we have some kind of insurmountable disagreement about what letters and words are. jp×g 09:26, 25 January 2023 (UTC)[reply]
@JPxG:

You've just fed me a variation of "you're not listening", with a little barb attached to the end. Really? That's who you are? I'm disappointed.

I read what you wrote, and I highly disagree with what you are saying...

You are saying that current copyright policy is enough: it prohibits copyrighted works from being posted to Wikipedia without the permission of the copyright holder, and that it is up to the editor to make sure that the material does not violate anyone's copyrights or Wikipedia's copyright policies.

My positions are...

1) that black box chatbots pose the danger of luring editors into violating copyright policy, that we may be faced with a deluge of copyright-violating derivative material because of it, and that some additional guidance would be appropriate: Like avoiding black box chatbots in favor of transparent ones, and...

2) that the copyrights to the natural language output composed by chatbots is unclear — what is clear is that the editor didn't write it. Since the editor didn't write it, does that mean that the editor does not own the copyrights to it? And if editors don't own the copyrights, should they be giving it to Wikipedia? Wikipedia should form a stance on the copyrights of chatbot-generated-output and present editors with guidance on this issue as well.

You have apparently been avoiding replying to those positions, and so my guess is that you are opposed to them. I strongly oppose the let's-stick-our-heads-in-the-sand approach that you support.    — The Transhumanist   10:55, 25 January 2023 (UTC)[reply]
P.S.: I don't think anything in writing is going to be enough. I expect that it will take software programming to deal with the problems Wikipedia will be subjected to by chatbot compositions. And that is beyond the scope of this venue. ;)    — The Transhumanist   11:31, 25 January 2023 (UTC)[reply]

To act or not to act

Like DALL-E last year, or NFTs the year before that. I'll believe it when I see it, and I can't see the value in spending even more time discussing a hypothetical future threat to Wikipedia. – Joe (talk) 08:39, 13 January 2023 (UTC)[reply]
The scariest yet most plausible thing is that this is happening with some of the articles but we aren't aware of it. I don't think raising awareness on this issue is a bad thing given how fast AI advances nowadays. CactiStaccingCrane (talk) 00:39, 14 January 2023 (UTC)[reply]
Sorry, I missed the "recent events". Where is the evidence for GPT problems on Wikipedia? —Kusma (talk) 11:57, 13 January 2023 (UTC)[reply]
Where's the evidence that Wikipedia can't cope with AI generated articles? doktorb wordsdeeds 14:07, 14 January 2023 (UTC)[reply]
@Doktorbuk: You are talking in terms of hindsight (asking to see what has already happened), rather than applying foresight to assess a potential threat by asking "What could happen?"

Here's an article from the New York Times -- imagine a similar effort directed at Wikipedia using thousands upon thousands of (seasoned) new accounts to support political POVs, revise history, censor opposing opinions, and spread other forms of misinformation:

https://www.nytimes.com/2023/01/15/opinion/ai-chatgpt-lobbying-democracy.html

It's only a matter of time before the powers that be shift their attention, and their tools, upon the English Wikipedia. The question is, are we ready for when we have to be? Here's an article that makes one wonder what these people will do now that they have Chat-GPT to work with:

https://www.theweek.in/news/world/2023/01/06/saudi-arabia-infiltrated-wikipedia-and-this-is-how-they-did-it.html

So, do we really need evidence that the English Wikipedia has already been breached by LLM-assisted POVers before proceeding? Or can we prepare for this in advance?    — The Transhumanist   00:34, 16 January 2023 (UTC)[reply]
Your reasoning seems to be that
  1. ChatGPT (or its equivalents) can write disinformation quickly and cheaply
  2. POV-pushers (governments, lobbies etc.) are currently limited by the time humans need to write disinformation
  3. Wikipedia is a prime target for such POV-pushers
  4. Therefore, ChatGPT (or its equivalents) will flood the gates, unless we do something.
I will grant you (1) is either already true or will likely be in the near future.
However, (2) is questionable (see that XKCD about old-fashioned human-crafted POV-pushing). I would guess coordinating the messaging and maintaining the disinformation is a much larger fraction of the costs than actually writing the text.
(3) is also dubious. Editing in a way that sticks is much harder on Wikipedia than in other places (such as facebook, reddit, etc.). Maybe it has more impact, but the cost-benefit analysis is not obvious.
Finally, inaction is always an option. It might not be a good option, it might even be the worst option, but it must be compared to other specific measures. "Something must be done" without specifics is just the politician's fallacy. In the absence of details about the threat, it’s hard to compare the possible countermeasures. TigraanClick here for my talk page ("private" contact) 16:41, 16 January 2023 (UTC)[reply]
@Tigraan, I think your list of assumptions is missing "5. People who want to corrupt Wikipedia (e.g., NPOV violations, stacking votes) can reasonably be expected to obey any prohibitions we announce on using this particular technology to achieve their illicit ends." WhatamIdoing (talk) 21:22, 17 January 2023 (UTC)[reply]
I wanted the list of assumptions to be a reasonable summary of (what I understand to be) TH’s argument; I suspect your suggestion is... not that. But I agree that’s part of the problem (which my last paragraph covers). TigraanClick here for my talk page ("private" contact) 10:37, 18 January 2023 (UTC)[reply]
Dear @Tigraan: I am concerned with a flood of WP-incompatible chatbot-generated content, whether by POV'ers or good-faith editors. But it won't be at any gates. The water-level will simply rise. If and when flooding begins, it will be a matter of bailing out the excess. There are three questions relevant to such potential flooding:

1) Will chatbots be designed in such a way to prevent flooding (and bailing) in the first place by minimizing inappropriate (unsourced, misinforming) content?

2) Will the bailing be automated?

3) Shall we wait to work on #1 & #2 until after flooding has begun, or prepare in advance?

Some editors seem doubtful that the addition of content generated by LLMs to Wikipedia beyond the manual capacity of our editors to process it will happen. And I don't know if it will happen, either. But, there is a continuous stream of strong indications that LLM-based tools will become ubiquitous in the not too distant future, for general use, which, by extension, includes using them to add content to Wikipedia. Here's another:
Google Calls In Help From Larry Page and Sergey Brin for A.I. Fight — New York Times]
And the technology push isn't limited to OpenAI and Google. Here's a search engine that uses a natural-language interface in both its queries and its answers:
Perplexity AI: Ask Anything
It is looking pretty clear that some major changes are on the horizon in the way computer users will be composing web content. It is also profoundly obvious that Wikipedia isn't ready right now for much more than the current volume of content creation that it is already handling. Maybe the volume won't increase by much, or maybe it will.

However, some editors are taking the potentiality that it will seriously, and it'll be interesting so see if their preparation efforts will be sufficient to stem the tide, if or when the tide rises. Sincerely,    — The Transhumanist   22:51, 20 January 2023 (UTC)[reply]
@The Transhumanist, I'm not sure that these questions are really suitable to a discussion on Wikipedia. The first one, for example: Will chatbots be designed in such a way to prevent flooding (and bailing) in the first place by minimizing inappropriate (unsourced, misinforming) content?
I'd re-phrase it like this:
"Will all of the people who are not us, including those who don't care about us, carefully design their software in such a way to be convenient for us?"
Answer: No. Or, at least, it is highly unreasonable to assume that the answer is yes for all of the people who write this sort of software, and it only takes one to risk a problem. WhatamIdoing (talk) 21:21, 23 January 2023 (UTC)[reply]
@WhatamIdoing:

That would more likely be true if members of WP or the WMF did not contact them. They are not producing these things in a vacuum. WP/WMF has a good relationship with Google, for example, which applies Wikipedia content extensively. It may be time to reach out to the companies developing chatbots too.

On the bright side, there's pressure coming from abroad, in the critique of chatbots, to be less "black box" and to provide references, which is one of the features that would help avoid problems.

Perplexity.ai already provides sources, which helps with verification efforts, and to see which ones are and are not from Wikipedia. Though, Perplexity.ai does not provide quote marks around passages that it quotes, and that is another problem. So, I guess they need to be contacted as well.

It looks very likely that chatbots will be used to compose content for other websites besides Wikipedia, and that their webpages may be included in chatbot training data too -- making an error-magnifying feedback loop a potentially huge problem for the chatbots. Too big to go unnoticed, hopefully.

It's important that we are aware of these issues if we are to have any chance in influencing solutions. Who knows, the chatbots, and/or the chatbot developers, may actually read this discussion. ;)

The WMF has been made aware of this discussion, so they can read it to prepare for discussions with participants in the chatbot sector. So, it is important that we get our concerns, and thoughts on design and strategy, in print.    — The Transhumanist   08:20, 25 January 2023 (UTC)[reply]
You are assuming that LLMs are being developed by a manageable number of identifiable companies, and hoping that all of them would like to protect Wikipedia.
But let's consider it from a different POV. Imagine that chatbot software is an open-source project, like Wikipedia. You have the Wikipedia:Right to fork open source projects – not just Wikipedia, but any open-source project. Anyone can add or subtract anything on their own setup. For example, if someone adds a "protect Wikipedia" module, then the next person could remove that, or even add a "scam Wikipedia" module.
I believe there will be some organizations who find that protecting Wikipedia aligns with their interests, and they will do so. But there will also be some organizations who find that protecting Wikipedia is exactly the opposite of their interests, e.g., content farms that hope they'll be cited as sources here so that their ad-filled webpages will get more traffic, and WP:UPE scammers who are hoping to reduce their costs by having their secret, internal-use-only chatbot write Wikipedia articles for clients of dubious notability, rather than paying a human to do that. I don't think that we can identify such actors, and I don't think they would change their behavior even if we talked to them.
On a tangent, the call for chatbots to cite sources and add quotation marks is probably based on a misunderstanding. LLMs aren't "quoting sources". They're predicting what a typical way to complete a sentence might be. If it spits out "The journey of a thousand miles begins with one step", it's not quoting Lao Tzu; it's saying "When I look in my database, and I see phrases that start with 'The journey of', the next bit is usually either 'a thousand miles' or 'a lifetime'. I'll pick one and see what comes next." WhatamIdoing (talk) 17:04, 25 January 2023 (UTC)[reply]
@WhatamIdoing: Good point concerning the ecosystem of chatbot developers - I was only considering the big name actors (Google, etc.), but anyone and their uncle can get involved. You are right, bad actors are inevitable and perhaps even rampant. Yet, the vast majority of chatbot use will likely be of the big name models (ChatGPT, etc.). So, contacting and working with them would be beneficial.

As for quoting, I have found that the AI search engine perplexity.ai, which includes inline source references in its natural language answers to users' questions, integrates passages verbatim from the referenced webpages into its answers without using quotation marks.    — The Transhumanist   09:32, 26 January 2023 (UTC)[reply]
I wonder what the Ask Jeeves developers are thinking about that. Perhaps they were just 20 years too soon. WhatamIdoing (talk) 17:46, 26 January 2023 (UTC)[reply]
That’s a blast from the past, Doug Weller talk 18:30, 26 January 2023 (UTC)[reply]
Looks like some spammers and malware distributors have embraced this technology:
WhatamIdoing (talk) 01:09, 28 January 2023 (UTC)[reply]
The second source above says "Likewise, anyone who uses the web to spread scams, fake news or misinformation in general may have an interest in a tool that creates credible, possibly even compelling, text at super-human speeds." We need detection tools, and fast. The "super-human speed" part could be a dead giveaway.    — The Transhumanist   10:00, 28 January 2023 (UTC)[reply]

Identifying chatbot-generated text

Zero0000's post is a good start. A simple way to crystalize the situation is to ask the human editor for their rationale for a particular phrase. North8000 (talk) 03:55, 14 January 2023 (UTC)[reply]

I feel like additions of large, overly-verbose unsourced text are something of a giveaway. See, for example, the first revision of Artwork title, written by ChatGPT. — Qwerfjkltalk 11:06, 14 January 2023 (UTC)[reply]
You can be confident that any giveaways are on the chatbot writer's list of things to fix in the next generation. Zerotalk 11:56, 14 January 2023 (UTC)[reply]
They should also fix the problem of the chatbots making stuff up. Someone should ask the chatbot writers to turn off the poetry and fiction generation algorithms, and any other algorithms that make things up, when the chatbots are composing expository text. Or add new algorithms to handle expository writing. Just the facts. And sources.    — The Transhumanist   00:42, 16 January 2023 (UTC)[reply]

Nature just published a piece about use of ChatGPT in scientific articles. Zerotalk 01:18, 21 January 2023 (UTC)[reply]

Is someone clear on what the copyright status of texts produced by LLMs is? From what I get, they may be considered derivative works from the dataset they were trained on. From [2]: As a result of the human authorship standard, “under U.S. current law, an AI-created work is likely either (1) a public domain work immediately upon creation and without a copyright owner capable of asserting rights or (2) a derivative work of the materials the AI tool was exposed to during training,” Esquenet continues. “Who owns the rights in such a derivative would likely be dependent on various issues, including where the dataset for training the AI tool originated, who, if anyone, owns the training dataset (or its individual components), and the level of similarity between any particular work in the training set and the AI work.” If they are derivative works then they cannot be published on Wikipedia just like this. Do we have more information on this? For example, does OpenAI specify somewhere the copyright status of the text produced by ChatGPT? Phlsph7 (talk) 09:09, 16 January 2023 (UTC)[reply]

The first question is whether a generated text that closely resembles an item from the training set is copyright infringement of that item. For instance, Microsoft Copilot happily outputs the Fast inverse square root code. I would expect that courts will judge such things to be copyright infringement. Copyright infringement statutes do not require to prove that the infringer copied a specific source (that would be difficult to prove), just that the content is substantially similar. Therefore, whether the tool is a simple ctrl-C ctrl-V or a sophisticated machine learning model should not make much difference.
The second question is whether OpenAI (or any other AI tool provider) can assert copyright on whatever the tools they provide create. The OpenAI terms of use seem relatively permissive, but others might be less generous. I do not know the answer to that question. I would hope they cannot, since they only provide tools (Microsoft should not be able to assert copyright on the text I write using Word, or the images I draw using Paint).
The third is whether a human using ChatGPT can assert copyright on ChatGPT answers, or otherwise constrain the use of the resulting text. The quote you give is probably based on the US copyright office’s position (taken during the monkey selfie copyright dispute): Because copyright law is limited to 'original intellectual conceptions of the author', the [copyright] office will refuse to register a claim if it determines that a human being did not create the work. However, giving a prompt to ChatGPT might or might not constitute significant creative input. The position that anything edited by a machine becomes public-domain is untenable (if I use an orthographic corrector on the draft of my novel, it does not turn it into PD), so it must be a question of degree. Also, non-US courts might have different opinions. TigraanClick here for my talk page ("private" contact) 16:06, 16 January 2023 (UTC)[reply]
As I understand it, the degree of access by the alleged infringer to the source text in question is a factor in determining infringement. Only a specific expression is protected by copyright; if you and I independently write the same sentence, one is not a copyright violation of the other. The amount of similar text also plays a role, since the larger it is, it's more improbable that it was created without copying.
Facts and natural laws can't be copyrighted; this also covers algorithms (though a particular expression can be copyrighted). So I can't copyright a single instance of a Javascript for-loop and claim rights to all Javascript for-loops as derivative work. In cases where the learning model creator is explicitly providing its model for use as a tool, I think (disclaimer: not a legal opinion) it is reasonable for this to be the same as a work for hire. Thus if the result is eligible for a new copyright owner independent of any source texts, the tool user would be the owner. (If I use a spellchecker on the latest bestselling novel, the result is not eligible for a new copyright owner.)
To be really safe, we'd want language models trained on public domain text. But I think it could be argued with a really large model trained on, say (just drawing numbers out of air), hundreds of thousands of documents with thousands of independent authors, the resulting correlations can no longer be attributed to specific input text, for cases where the output is not a significantly long passage substantially similar to a specific source text. isaacl (talk) 18:05, 16 January 2023 (UTC)[reply]
One of the main issues to deal with would be the following: an editor tells ChatGPT to write a text on a topic and then adds this text in the form of a paragraph/section/article to Wikipedia and thereby publishes it under Creative Commons/GNU license. The question is: what are the chances that this constitutes some form of copyright violation? This might concern specifically problems with the 1st and the 2nd question addressed by Tigraan, i.e. whether the copyright of someone whose work was part of the training set was violated and whether openAI's copyright was violated. For the first question, it's probably relevant what the copyright status of the texts in the training set is and how similar the produced text is to the texts in the training set, as isaacl points out. Answering these questions would be quite relevant for any Wikipedia policy on the topic, like the one JPxG is currently drafting. Phlsph7 (talk) 07:18, 17 January 2023 (UTC)[reply]
  • With respect to the issue of whether LLM output inherently violates copyright law: the copyright status of LLM-generated text is not defined by statute, so it is hard to make confident claims, but precedent exists for computer-generated art and other works created by non-humans. Here is what the US Copyright office has to say:
"Compendium of U.S. Copyright Office Practices, § 313.2" (PDF). United States Copyright Office. 22 December 2014. p. 22. Retrieved 18 January 2023.
The Office will not register works produced by nature, animals, or plants. Likewise, the Office cannot register a work purportedly created by divine or supernatural beings, although the Office may register a work where the application or the deposit copy(ies)state that the work was inspired by a divine spirit.
[...]
Similarly, the Office will not register works produced by a machine or mere mechanical process that operates randomly or automatically without any creative input or intervention from a human author.
It's not quite clear to me what the applicability is in this case. On Commons, the template and copyright category for PD-algorithm asserts that all algorithmically-generated works are public domain ("This file is in the public domain because, as the work of a computer algorithm or artificial intelligence, it has no human author in whom copyright is vested"). Whether artificial neural networks are capable of producing original intellectual output is less of a legal issue and more of a philosophical/anthropological one. It should be noted that human brains are themselves neural networks; much has been said, in a variety of fields, on the subject of whether humans create original works versus whether they merely juxtapose or recombine motifs and concepts that they're exposed to through participation in society. While interesting (and humbling), these discussions are unrelated to whether neural networks which have been exposed to copyrighted material in the course of their existence are capable of later creating original works under the purview of intellectual property law: they do. If this was not the case, a large majority of creative work would be illegal (good luck finding a band where none of the musicians have ever heard a copyrighted song before).
It does seem obvious that, say, OpenAI does not hold a claim to copyright on all works produced using its products: Adobe does not own copyright on all works produced using Photoshop, and Apple does not own copyright on all photographs taken with iPhones. If they did, this would certainly be a bold claim which required some evidence beyond idiosyncratic interpretation of vague phrasing in a EULA somewhere.
With respect to whether LLM output is capable of violating copyright law, I think the answer is obviously "yes". Here is an example, which I have just prompted from ChatGPT:
What are the lyrics to "Moonage Daydream", by David Bowie?
Keep your 'lectric eye on me babe
Put your ray gun to my head
Press your space face close to mine, love
Freak out in a moonage daydream oh yeah
These are, indeed, the verbatim lyrics to an unambiguously copyrighted song. Of note is the fact that I specifically asked it to do this. Compare with this prompt, using the first couple line of the song:
I'm an alligator / I'm a mama papa comin' for you / I'm a space invader / I'll be a rock-'n-rollin' bitch for you
I'm a force of nature, a wild beast / A creature of the swamps, a king of the feast / I'm a living legend, a natural high / 'Cause I'm an alligator, and I'll never die
While this would be a sick Judas Priest song, it isn't one, and in fact, "I'm+an+alligator+and+I'll+never+die" this text seems to never have been published. jp×g 01:57, 19 January 2023 (UTC)[reply]
It bears mentioning that I, a human being, also know the lyrics to "Moonage Daydream", and am capable of typing them into the edit box with no assistance whatsoever; doing this in an article would still be a copyright violation. I think that, for LLMs, the critical piece of knowledge is that output needs to be checked, and that it is insufficient to assume that generated text is always novel. jp×g 02:15, 19 January 2023 (UTC)[reply]
On the narrow point of who owns ChatGPT outpt, Sec 3(a) of their ToS states "OpenAI hereby assigns to you all its right, title and interest in and to Output...You are responsible for Content, including for ensuring that it does not violate any applicable law or these Terms." I thought there was an additional requirement to state ChatGPt was used, but I do not see it in the terms. Slywriter (talk) 02:14, 19 January 2023 (UTC)[reply]
@JPxG: since you mention computer-generated art: there are some lawsuits against Stability AI based on the claim that they violated the copyrights of people whose images were used in the training set. See [3] and [4]. The case seems to be similar to LLMs, with the main difference being that their AI trains on images and creates images while LLMs train on text and create text.
If I interpret the statement by the US Copyright office correctly, it seems to claim that a person can't own the copyright of a work that was created by a random machine process without creative input. It does not say that such processes cannot violate someone else's copyright. This would be in tune with the lawsuits mentioned above.
I think it's also unlikely that every output is a copyright violation. For example, if you just give it a sentence and tell it to correct spelling mistakes, there should be no problem in using the output. Phlsph7 (talk) 06:33, 19 January 2023 (UTC)[reply]
@Slywriter: Their sharing policy demands that Indicate that the content is AI-generated in a way no user could reasonably miss or misunderstand. Phlsph7 (talk) 09:23, 19 January 2023 (UTC)[reply]
LLM-assisted edits need to be appropriately marked as such in the history. —Alalch E. 01:38, 21 January 2023 (UTC)[reply]
Yes, and I think that "in a way no user could reasonably miss or misunderstand" requires the use of a notice in the article itself as well. –dlthewave 13:53, 21 January 2023 (UTC)[reply]
Agreed. It's easy for the average reader to miss an edit summary in the article history. So in-text attribution may be required. Phlsph7 (talk) 08:54, 22 January 2023 (UTC)[reply]
@Slywriter, JPxG, Phlsph7, Alalch E., and Dlthewave: Concerning the TOS clause that states "OpenAI hereby assigns to you all its right, title and interest in and to Output...You are responsible for Content, including for ensuring that it does not violate any applicable law or these Terms." — does that mean that Chat-GPT cannot legally produce the exact same output twice without violating the right, title, and interest that it previously assigned?    — The Transhumanist   20:47, 21 January 2023 (UTC)[reply]
I'm not sure what leads you to that conclusion. The licence does not grant you exclusive use to anything. isaacl (talk) 22:09, 21 January 2023 (UTC)[reply]
@Isaacl: I didn't come to a conclusion, I just asked a question, pertaining to Sec 3(a) of their ToS as referred to and quoted by Slywriter above, and repeat quoted by me. It appears you missed the quote somehow, because you didn't comment on it. To what license are you referring, and what relation does it have to the passage we quoted from the TOS?    — The Transhumanist   02:52, 22 January 2023 (UTC)[reply]
I don't know why you would ask the question you did, since the section you quoted did not say anything about granting an exclusive right, title, and interest to any output. isaacl (talk) 03:08, 22 January 2023 (UTC)[reply]
I don't think this has something to do with producing the same output for different users. It should be easy to find mock queries to which it often responds with the same output, for example, by asking it to "Say the word 'Hello'" or for simple translations. Phlsph7 (talk) 06:50, 22 January 2023 (UTC)[reply]
@Isaacl: Well, I checked the section again, and it is right there in plain English. It uses the word "assigns" instead of "grants", and it says "all its" instead of "exclusive". So, once it "assigns all its right, title, and interest in and to Output", how can it legally ever produce that same output again? (Because it already assigned it away).    — The Transhumanist   09:20, 22 January 2023 (UTC)[reply]
As I understand it, anyone can assign all their rights to the output of ChatGPT to someone else. In a similar way, I could assign to you all my rights to the Harry Potter series. This would not be of much use to you since the expression "all my rights" just refers to "no rights" in this case. Phlsph7 (talk) 09:32, 22 January 2023 (UTC)[reply]
@Phlsph7: In section 3a of the TOS, it's OpenAI that is assigning its rights to the chatbot output generated for the user. If Chat-GPT writes you a 3 paragraph explanation of gravity, and OpenAI has assigned you its rights to that explanation, can Chat-GPT legally write that exact same output for somebody else?    — The Transhumanist   09:58, 22 January 2023 (UTC)[reply]
I assume it works something like the following: it depends on whether openAI had any copyrights on it in the first place. If it did then this may be a problem because creating the copy for the second user might violate the newly obtained copyright of the first user. If it didn't then it presumably wouldn't be a problem because assigning all its rights to the first user effectively didn't do anything. But I don't think that this particular issue is very relevant for Wikipedia. Phlsph7 (talk) 10:14, 22 January 2023 (UTC)[reply]
The tool made no guarantee that it wouldn't generate the same output again for another user. The tool is in essence passing any rights of ownership (if they exist) in the original input through to the output. isaacl (talk) 19:24, 22 January 2023 (UTC)[reply]

Testing Chat-GPT's algorithm

Does Chat-GPT produce the same output to the same prompt given to it by 2 or more different users? Do any two want to try that?    — The Transhumanist   20:54, 21 January 2023 (UTC)[reply]

@The Transhumanist, not necessarily. It has a "temperature" factor (randomness). — Qwerfjkltalk 21:01, 21 January 2023 (UTC)[reply]

@Phlsph7, Isaacl, Tigraan, Dlthewave, Slywriter, and JPxG:

I have found the this to be perplexing...

I entered the following prompt into Perplexity.ai:

who owns the copyrights of perplexity.ai's answers

And it returned the following answer:

"According to US copyright law, works generated solely by a machine are not eligible for copyright protection[1] . Therefore, the copyrights of Perplexity AI's answers belong to its co-founder and CEO, Aravind Srinivas[2]."

It looks like I just broke copyright by copying it here. But this contradicts the title of the first source provided (the second source is behind a registration wall):

The scary truth about AI copyright is nobody knows what will happen next — The Verge

Assuming that Aravind Srinivas does not own the output, I have some questions about posting AI search engine results: I look forward to your replies to the below questions.    — The Transhumanist   19:27, 21 January 2023 (UTC)[reply]

Question 1: Does putting quotation marks around a small copy/paste selection from an AI search engine, like the passage above, fall under fair use?

I would assume so. It's similar to how you can quote from copyrighted books. There are some limitations, for example, concerning the length of the cited text. And it should be clear where the quotation came from. Phlsph7 (talk) 06:57, 22 January 2023 (UTC)[reply]

Question 2: Would that protect Wikipedia (and its editors) from a derivative work claim?

From [5]: "In its most general sense, a fair use is any copying of copyrighted material done for a limited and “transformative” purpose, such as to comment upon, criticize, or parody a copyrighted work. Such uses can be done without permission from the copyright owner. In other words, fair use is a defense against a claim of copyright infringement. If your use qualifies as a fair use, then it would not be considered an infringement." Phlsph7 (talk) 07:06, 22 January 2023 (UTC)[reply]

In our discussion so far, we haven't been able to conclusively figure out whether someone owns the copyright at all and, if so, who. That 2 users get and use the same response would be just a special case. Phlsph7 (talk) 07:14, 22 January 2023 (UTC)[reply]

Question 4: Would running a passage (from a chatty AI search engine) through a plagiarism checker be enough, before copying it into Wikipedia?

Plagiarism checkers are not perfect so they can't ensure that no plagiarism/copyright infringement was committed. The question would be whether they are good enough for our purposes, i.e. whether they are quite reliable for spotting plagiarism/copyright infringement pertaining to AI-generated texts. Phlsph7 (talk) 07:26, 22 January 2023 (UTC)[reply]

Question 5: Does Wikipedia policy allow an editor to click "Publish changes" for content that the editor did not personally compose?

Clarification: Clicking "Publish changes" implies that the editor composed the changes. Can an editor publish changes that they did not personally compose, that were composed by a chatbot search engine? (Please quote and provide links to the specific policies that allow or disallow this). Thank you.    — The Transhumanist   20:13, 21 January 2023 (UTC)[reply]

That would probably be a case of WP:PLAGIARISM even if no copyright infringement is involved. According to the summary: "Do not make the work of others look like your own. Give credit where it is due." Phlsph7 (talk) 07:06, 22 January 2023 (UTC)[reply]
It would be similar to copying public domain/open license content to Wikipedia, no? This is covered by several guidelines and explainers such as WP:FREECOPY and Help:Adding open license text to Wikipedia. As long as there's proper attribution, there's no general expectation that editors must compose the text themselves. –dlthewave 13:22, 22 January 2023 (UTC)[reply]
@Dlthewave and Phlsph7: Interesting. So, if you prompted a chatbot to write a new paragraph for the article on cream cheese and you add that to the article, you include an attribution to the chatbot in the edit summary? What do you put in the source reference?    — The Transhumanist   11:58, 25 January 2023 (UTC)[reply]
@The Transhumanist: As I understand it, for WP:PLAGIARISM it's sufficient to mention the source in the edit summary. You would have to find and add other reliable sources yourself since ChatGPT provides no sources or sometimes invents non-existing sources. However, for the Sharing & Publication Policy of openAI, in-text attribution would probably be necessary. So to comply with it, you would have to start the paragraph on cream cheese with something like "According to ChatGPT,...". This way, the text couldn't be used at all since ChatGPT is not a reliable source. Phlsph7 (talk) 12:23, 25 January 2023 (UTC)[reply]

First ANI case

Just a head up, a thread in Wikipedia:Administrators' noticeboard/Incidents has just opened about an user abusing AI-generated content at Wikipedia:Administrators'_noticeboard/Incidents#Artificial-Info22_using_AI_to_produce_articles. Sure, the editor in question did not made an edit in the mainspace, but the fact that this is happening at ANI is pretty concerning in its own right. I afraid that someone may have covertly spam articles with AI text already. CactiStaccingCrane 15:31, 23 January 2023 (UTC)[reply]

I am now adding the {{disputed}} template when encountering an AI-generated article, based on the following from the ChatGPT FAQ: These models were trained on vast amounts of data from the internet written by humans, including conversations, so the responses it provides may sound human-like. It is important to keep in mind that this is a direct result of the system's design (i.e. maximizing the similarity between outputs and the dataset the models were trained on) and that such outputs may be inaccurate, untruthful, and otherwise misleading at times. The commonality of all the AI-generated articles I've encountered so far (4, to be honest) is that they are not properly footnoted, implying that the author has not confirmed that the AI output is correct. The disputed tag seems to cover this issue well. I'm also dropping a note on the article's talk page explaining the link between AI output and correctness. — rsjaffe 🗣️ 01:00, 24 January 2023 (UTC)[reply]
ANI case is wrapping up. The first three articles were written by a hoaxer, and the refs for two of the articles may have been generated as well. The fourth article was promoting a company. — rsjaffe 🗣️ 04:18, 24 January 2023 (UTC)[reply]
This gives us a good look at the type of plausible-sounding nonsense that we can expect from LLM output. Comparing the draft (archive version, since it will likely be deleted soon) to Gecko, I see a few factual errors right away:
  • Not all geckoes belong to the family Gekkonidae, which doesn't have 1500 species.
  • Not all geckos have specialized toe pads that allow them to climb vertical surfaces.
  • The largest geckos are 23"-24", not 10".
  • Not all geckos are oviparous; some bear live young.
When this type of content is submitted, it needs to be thrown out straightaway. –dlthewave 17:09, 24 January 2023 (UTC)[reply]

More detected AI-generated papers

I started screening Drafts more carefully and am getting a number of hits corresponding to probable AI-generated articles (or at least part of the article is AI-generated). Examples include:

The list could go on, but I think this is enough to see some information about this. These pages tend to be created by users with few edits. A number of users are doing this, not just one or two. Conclusion: the tsunami has arrived. — rsjaffe 🗣️ 03:15, 25 January 2023 (UTC)[reply]

I think that mw:ORES (used in recent changes to highlight bad faith and vandalism) should integrate a screening mechanism for GPT-3 and other bots asap. I envision this is already a huge problem when large amount of hoaxes can be disguised as good content and we wouldn't even know about it. CactiStaccingCrane 03:21, 25 January 2023 (UTC)[reply]
Looking at the first few, the sourcing is not up to the standards of promotion to article space. Once clearly bad sources are removed and unsourced claims are tagged, this can clearly be seen. If AI ever gets to the point of being able to write an article that provides accurate information properly and verifiably sourced to reliable sources, then I'll be happy to have it writing for us. BD2412 T 03:46, 25 January 2023 (UTC)[reply]
@CactiStaccingCrane: It might be a good idea, but in reality, it requires a lot of effort from WMF to integrate openai-detector into mw:ORES. And I agree with @BD2412 for pointing out some drafts are promotional, which I think is a problem even before ChatGPT or even GPT-3 exist. 2001:448A:304F:52BA:8D12:5E35:69B7:8E09 (talk) 03:50, 25 January 2023 (UTC)[reply]
Well, some AI-generated articles have made it into article space. The first one I found has lots of text that's probably not AI-generated, but has a big hunk that is. Pavilion of Harmony, from "The Harmony Pavilion of New Asia College..." to "unique addition to the campus of New Asia College.", after removing the footnote indicators that confuse the analysis, rates as 99.98% fake. So the problem will leak into article space. And this means we need a way to pay special scrutiny to the AI-generated section, as that section is likely to have plausible but false information, given the way current AI models work. — rsjaffe 🗣️ 04:13, 25 January 2023 (UTC)[reply]
I agree that this could easily turn into a major problem as LLMs become more popular. As discussed at #Copyright status, these drafts violate at least WP:PLAGIARISM but probably also the Sharing & Publication Policy of openAI (if they were created using openAI tools). If AI-detectors are reliable, including them in mw:ORES would probably help a lot to mitigate the problem in case such an integration is feasible. Another alternative would be to create a bot that checks new submissions and tags them if they score a high value. A further thing to do at some point might be to make the editors reviewing drafts and new articles aware of this problem. Phlsph7 (talk) 06:28, 25 January 2023 (UTC)[reply]
10 Best ChatGPT Chrome Extensions You Need to Check Out Doug Weller talk 10:06, 25 January 2023 (UTC)[reply]
@Phlsph7 said: A further thing to do at some point might be to make the editors reviewing drafts and new articles aware of this problem. That's how I stumbled unsuspectingly upon this issue. I'm a new page patroller. I think they need to be looped in now, as that is the only guaranteed review step for new articles, and LLM-generated articles are already appearing. (I'm hoping that those users allowed to have their articles bypass this process won't abuse LLMs.) — rsjaffe 🗣️ 17:09, 25 January 2023 (UTC)[reply]
@Rsjaffe: Wikipedia:New_pages_patrol has various instructions on the different issues that new page patrollers need to be aware of. Maybe somewhere in there, a subsection could be added on AI-generated articles. Among other things, it should give a short explanation of what it is (the user tells the AI to generate an article in a matter of seconds and copy-pastes the results), what the problems are (plagiarism, false statements, no or invented sources, possibly copyright violation), and how to spot them (things AI-generated articles have in common and tools to detect them, like https://openai-openai-detector.hf.space/). Phlsph7 (talk) 19:31, 25 January 2023 (UTC)[reply]
Can I recommend, at least for right now, that some of these pages be copy-pasted into projectspace somewhere, so that we can see what they actually look like? I feel like these discussions basically have a couple-hour-long window outside of which it's impossible to see what everyone is talking about. jp×g 19:12, 25 January 2023 (UTC)[reply]
Some trials on articles I've written (using the lead paragraphs): Powder House Island and Nina Jankowicz are estimated at 99% human-written, but First Wikipedia edit is at 20% GPT for some reason. 1-Pentadecanol returns 40% GPT based on the first sentence, which decreases with subsequent sentences to 99%. However, when I asked ChatGPT to "write me a Wikipedia article about 1-pentadecanol", the result (which is viewable as an HTML comment if you edit this section) was estimated as 92% human-written. I don't know exactly what the underlying mechanism of this tool is, but we may want to take its output with a grain of salt. jp×g 19:23, 25 January 2023 (UTC)[reply]
The most recent articles I pulled (the ones in this section) were all initially detected by me using a search for a "signature" of a particular type of way of producing text for a particular version of LLM engine (I'm being vague as to not give hints to those who are trying to evade detection). I then visually confirm the signature. Then I run it through the GPT-2 detector. And then I'm only listing pages with > 99% chance of being LLM-generated. I'm 100% sure that the ones I've listed above are LLM-generated, but I'm also certain that this is only detecting that subset being generated under that specific set of conditions. — rsjaffe 🗣️ 19:24, 25 January 2023 (UTC)[reply]
Also, you have to remove the bracketed reference numbers (e.g., [2]) from the text as well as any intercalated headings to give an accurate score. — rsjaffe 🗣️ 19:26, 25 January 2023 (UTC)[reply]
To give an example of some LLM-looking text that doesn't score high on the GPT-2 detector, look at, ironically enough, Draft:Large Language Model. It scores 99.52% real, but sure looks fake to me. — rsjaffe 🗣️ 19:30, 25 January 2023 (UTC)[reply]
That detector works great if a text is modified, but it will crash when two or three words are modified. CactiStaccingCrane 19:32, 25 January 2023 (UTC)[reply]
Here are a few more detectors:
To test them, I asked ChatGPT to "write a Wikipedia article on cream cheese". It produces the following:
Extended content
Cream cheese is a soft, mild-tasting, white cheese that is made from a mixture of cream and milk. It is typically sold in blocks, and is commonly used as a spread on bread or crackers, as well as in cooking and baking.
The origins of cream cheese can be traced back to the early 19th century in the United States, where it was first produced by William Lawrence, a farmer and cheese maker from New York. Lawrence accidentally discovered the process for making cream cheese while trying to replicate the French cheese Neufchâtel. He began producing and selling the cheese under the brand name "Philadelphia Cream Cheese."
Cream cheese is made by blending cream and milk, and then adding a starter culture and rennet to coagulate the mixture. The curds are then drained and the remaining whey is discarded. The curds are then salted and packaged.
Cream cheese is a versatile ingredient and can be used in a variety of dishes, including dips, spreads, and baked goods. It is a key ingredient in many traditional American dishes, such as cheesecake, bagels with cream cheese, and lox and cream cheese. It is also commonly used in cooking and baking as a replacement for butter or other fats, as it adds a rich, creamy texture to dishes.
Cream cheese is also a popular ingredient in many international cuisines, particularly in Asian cooking. In China, cream cheese is often used as a filling in steamed buns and dumplings, while in Japan, it is used in sweet and savory dishes, such as cream cheese-filled pastries.
Cream cheese is typically sold in blocks, and can be found in most grocery stores and supermarkets. It is also available in a variety of flavors, including plain, herb, and smoked salmon.
In conclusion, Cream Cheese is a soft, mild-tasting white cheese made from a mixture of cream and milk, it is a versatile ingredient used in many traditional American and international dishes and it's also commonly used in cooking and baking as a replacement for butter or other fats. It is typically sold in blocks and is available in a variety of flavors.
All the detectors agree that this text is AI-generated. When I fed them with the lead of the article Wikipedia (pure text without reference signs), they all agree that it's human-generated. Phlsph7 (talk) 20:03, 25 January 2023 (UTC)[reply]
Some attempt at figuring out what these are: I think that they are all based on the same code from HuggingFace. The actual source is here. I think that it may be simple enough for me to deploy it on a Toolforge test account; some of these websites seem unbelievably seedy. For example, "Content At Scale" advertises:
Want undetectable AI content? Our platform is the only one of it's kind that allows you to upload up to 100 keywords and get back 100 entire human quality blog posts (title to conclusion) without any human intervention. All the while, bypassing AI detection as it's the most human-like AI content ever produced. Our proprietary system uses a mix of 3 AI engines, NLP and semantic analysis algorithms, crawls Google, and parses all the top ranking content to put it all together. This isn't an AI writing assistant, this is a human level long-form blog post producing machine!
Certainly seedy if they are making it up... but also very seedy if it's true! jp×g 23:17, 25 January 2023 (UTC)[reply]

Quoting a chatbot

Below are some verification-related questions pertaining to chatbots.    — The Transhumanist   12:10, 26 January 2023 (UTC)[reply]

How would pasting in content generated by a chatbot be interpreted under WP:VER's requirement that all quotes must be referenced?

WP:VER states that all quotes must be supported by inline citations. If the chatbot's text is unique rather than preexisting somewhere else, using it would in essence be quoting the chatbot — how could that, as it isn't recorded anywhere, be referenced for verification purposes?

For not requiring referencing the quote of a chatbot, would WP:VER need to be modified?

News update

We need to get and stay ahead of this AI thing. See the following to get an idea how fast this movement is progressing:

  1. ⭕ What People Are Missing About Microsoft’s $10B Investment In OpenAI : GPT3
  2. Travis Tang on LinkedIn: ChatGPT for Data Science Prompts - 60 examples of what it can do
  3. How to write an effective GPT-3 prompt | Zapier
  4. OpenAI Licenses GPT-3 Technology to Microsoft (not exclusive)
  5. OpenAI's investments
  6. Should ChatGPT be used to write Wikipedia articles?
This article features the following Wikipedia article, initially composed using Chat-GPT: Artwork title, by Pharos, a great example of how a chatbot can be used by a responsible editor. Maybe a blanket ban is too much, and guidelines on how to use it correctly would be better. Makes auto-removal harder, though.
See an explanation and its discussions here: Talk:Artwork title

I hope you find these articles informative. Feel free to post more links, and comments, below.    — The Transhumanist   16:05, 28 January 2023 (UTC)[reply]

Crystallize chatbot discussions into a policy?

I think that there is a long list of things that make chatbot content objectively bad for / incompatible with Wikipedia in its current form and methods. Without elaborating here, a few quick notes are the inherent "blackbox" nature of AI, the inherent unaccountability for content, the inherent non-linking of content to sourcing, the fact that they do not know or follow Wikipedia policies and guidelines, (which themselves are a fuzzy ecosystem rather than categorical rules) they do know take into account what is already in the article, they (as a practical matter) remove accountability and expectations from the person who added the material. They also would destroy the ability to obtain volunteer time to review what has been put in. Most people willing to spend time to review something because they know that a human editor has take the time to write it would not be willing to spend large amounts of time dealing with something generated by a bot in a few seconds.

My thought is that we should say that such chatbot generated content is not allowed in Wikipedia. This is just briefly written, I or someone could flesh this out into something carefully written if there is interest.

We can and should decide this without or prior to solving the question of how to detect and enforce. A premise of having to solve detection and enforcement before step #1 would be a poison pill for accomplishing step one. Sincerely,North8000 (talk) 20:42, 25 January 2023 (UTC)[reply]

@North8000: There has been a draft guideline at Wikipedia:Large language models for a few weeks. I do not know that banning their use entirely is a good idea, but it seems quite obvious that just copy-pasting gigantic chunks of text directly from the model's output into the edit box is not a good idea (and almost zero percent likely to result in usable articles). I will try to write some stronger wording emphasizing that nobody should be doing this. jp×g 22:46, 25 January 2023 (UTC)[reply]
  • Support blanket ban - I agree with North8000 that the policy should be "Chatbot generated content is not allowed in Wikipedia." I think this should apply to all namespaces, including talk pages. As chatbots improve, the policy can be changed, but right now, chatbot use for generating WP content appears to be a can of worms.    — The Transhumanist   01:03, 26 January 2023 (UTC)[reply]
    Pertaining to talk pages and forums, I meant not allowing the use of chatbots to generate a user's statements in a discussion. Posting chatbot output on a talk or forum page as an example in order to discuss it, is appropriate.    — The Transhumanist   12:33, 26 January 2023 (UTC)[reply]
    Support auto-screening - Pharos and their article Artwork title changed my mind (and is featured in Should ChatGPT be used to write Wikipedia articles?). I'd like to see a draft on guidance for the responsible use of chatbots in writing articles, including Pharos' approach. Meanwhile, our tech persons can work on automating the removal of undisclosed chatbot additions and the tagging and eventual removal of other entries that don't get edited within a reasonable time frame, or that are part of a pattern of posting disclosed but unedited chatbot submissions. Donald Albury was right, bad actors are going to spam Wikipedia with chatbot crap whether we ban it or not. Therefore, we should allow good actors to help offset their impact. Which brings us to the rest of the Web: it will be subject to hosting chatbot content, and so, as we are Wikipedia, we should trailblaze how to do it right.   — The Transhumanist   16:21, 28 January 2023 (UTC)[reply]
  • Oppose a blanket ban. I think JPxG's demonstration shows that a LLM can be good for repetitive tasks like formatting a table, as long as a human validates the output. Actual generated prose is likely to be unsourced or sourced to fake sources, and so is already covered by existing policies. -- King of ♥ 01:23, 26 January 2023 (UTC)[reply]
  • Oppose any blanket ban. Do think CSD modifications are needed to quickly remove algorithmically generated articles (AI is market speak here, it is not sentient and intelligence is debatable) and some formal guidance for editors would be useful. It's a tool like AWB, Twinkle, and any other scripting used. Used properly, it can cut down tedious work Slywriter (talk) 01:30, 26 January 2023 (UTC)[reply]
  • Support blanket ban with the possibility of allowing specific use cases as we learn more. Our guiding principle should be that AI is completely inappropriate for creating or editing prose. Given the amount of plausible-sounding nonsense we've seen in recent AI-generated drafts, I also don't trust it for coding work such as formatting text or rotating tables until its reliability has been demonstrated for the specific task. This should apply to article, talk and draft spaces with very limited exceptions for demonstration purposes. –dlthewave 03:24, 26 January 2023 (UTC)[reply]
With respect to tables and templates, what do you envision as "demonstrating reliability"? It is not exactly brain surgery to look at a table and see if the columns or the parameters or whatever are in the right places. You have to do this anyway: we currently do not require editors to prove that they have never typed an extra } and had to go back and fix it. jp×g 05:35, 26 January 2023 (UTC)[reply]
  • Oppose for a variety of reasons, as discussed by others above. But I'll go with the most basic one: how would any of this be actionable? I know you claim it's a poison pill problem to raise, but there's no way whatsoever to detect this with certainty. We'd be using an outside tool to claim text is AI written and then delete things based on that claim. I don't care how many 9's you've got in the decimal places, there's no way to be infallible here. If the editor that added the text says they wrote it themselves, are we just going to say that they're lying and that they have to re-write it or something? There's not even evidence of copyvio in such a case and if the added content meets all other requirements, including proper verifiable sourcing, then I see no way to enforce such a ban. SilverserenC 03:33, 26 January 2023 (UTC)[reply]
    If the detector is good, one mistaken evaluation is possible, but a consistent output of one or the other is solid. Dege31 (talk) 17:49, 26 January 2023 (UTC)[reply]
  • Comment I'm not fully decided whether banning it entirely is the right course of action at this stage. But there is a significant potential to abuse it so most forms of non-trivial AI-assisted edits should be strongly discouraged. The policy should make it very clear that any addition of AI-generated text needs to be labeled as such in the edit summary to avoid WP:PLAGIARISM. The editors also need to be reminded to obey the license and sharing policy of the AI provider. In the case of ChatGPT, for example, in-text attribution is apparently required. Phlsph7 (talk) 08:53, 26 January 2023 (UTC)[reply]
  • Support blanket ban of any LLM-generated text. Not sure yet about using such tools purely for lay-out, but it should not be allowed for either generating new text or for rephrasing existing text, as both cases are way too problematic. As for "how is this enforceable", just like other difficult policies, where near certainty is sufficient (like WP:DUCK for socks, which isn't infallible but good enough). Advantages of a policy are also e.g. when a newbie says something like "why was my article deleted, it was generated by ChatGTP so has to be good", one can easily point to the policy to explain that it isn't allowed instead of having this discussion again and again. Fram (talk) 09:04, 26 January 2023 (UTC)[reply]
    That's a good point about potential problems with enforcing it. Having a policy can be useful to discourage certain types of behavior even if it is difficult to enforce in every case. We'll have to see how useful and reliable AI-detectors are in this process. Phlsph7 (talk) 09:15, 26 January 2023 (UTC)[reply]
  • I think the way forward is even stronger expectations on sourcing. If you can't provide the content of the reliable source for your edit, it should be reverted. (This would include people machine translating foreign Wikipedia articles without having access to the original sources). —Kusma (talk) 10:30, 26 January 2023 (UTC)[reply]
  • I assume that you are referring to automatic reversion (correct me if I'm incorrect). What about the wording of the verification policy that reads "challenged or likely to be challenged"? If it is not challenged or likely to be challenged, it doesn't need references. How will a bot be able to tell the difference between what does and does not require references? Or would the bot's removal of an edit constitute a challenge? Whether reversion is automated or not, should all new content to Wikipedia be challenged by default? That would require a change to WP:V, and that seems unlikely to happen.    — The Transhumanist   11:36, 26 January 2023 (UTC)[reply]
I'm not talking about automatic reversion, can't see how to do that without AI. And the verification policy is applied differently to new and existing articles; for new articles, we are in practice already expecting much better sourcing than "likely to be challenged" (just look at what will be rejected by AFC). Perhaps we should expand this to addition of content to existing articles. —Kusma (talk) 13:56, 26 January 2023 (UTC)[reply]
  • Support blanket ban for now; would be fine having a discussion later on allowable use cases, but I'd rather we started with a total blanket ban first, and then itemize specific possible use cases if we later decide there's some utility. --Jayron32 12:47, 26 January 2023 (UTC)[reply]
  • Support blanket ban for prose as there is significant copyright violation concerns about AI text. Text generated by bot will not substitute reliable sourcing. Maybe in the future when Abstract Wikipedia come online, we can give some leeway for bots to generate text based on reliably cited info, but for now, it's just too risky for the project. CactiStaccingCrane 13:34, 26 January 2023 (UTC)[reply]
If there is interest in at least reviewing this possibility (which it sounds like there is), as noted there it needs to be written better than my initial trail balloon above. I'll do that but still keep it short. I think that it can be done in a way that deals with the main enforcability questions and also allows described useful uses by allowing bot-assisted editor generated content. I'll do that within a 1/2 day. North8000 (talk) 14:33, 26 January 2023 (UTC)[reply]
  • support ban for text additionsThis is so susceptible to abuse and inserting misleading content that it should be banned. There is little benefit of allowing text generation and much harm.
— rsjaffe 🗣️ 16:05, 26 January 2023 (UTC)[reply]
  • Oppose new policy that goes beyond small clarifications of WP:BOTPOL. I think it is obvious that ChatGPT is already covered by it and I do not see what modifications are proposed.
Things that go against a content policy (WP:COPYVIO, WP:V, etc.) should be reverted / deleted on those grounds, and on those grounds alone; editors that make many such edits should be warned and then blocked. Editors who make faster edits than reasonably possible by hand should be dealt according to WP:MEATBOT.
I oppose any policy to revert / delete / ban based solely on a "seems bot-written" criterion, unless and until it has been proven that (1) this is a real, time-consuming problem on Wikipedia, and not a few random tests within the sea of vandalism, and (2) whatever criterion is used has been independently tested to establish its sensitivity and specificity and validated by the community. TigraanClick here for my talk page ("private" contact) 17:24, 26 January 2023 (UTC)[reply]
I also thought it obviously fell under BOTPOL, but Wikipedia:Bots/Noticeboard/Archive_17#Do_large_language_models_and_chatbots_(like_ChatGPT)_fall_under_the_bot_policy? this discussion shows some uncertainty. –dlthewave 20:04, 26 January 2023 (UTC)[reply]
  • Oppose a blanket ban. I agree in general with the oppose reasons given above. I also think such a ban would be "virtue signaling" without being effective. Editors who want to add AI-generated material to Wikipedia will not be stopped by such a policy. Consider how often our existing policies stop editors from adding un-verifiable, POV-pushing content. What we can do is work on effective strategies for detecting and removing un-sourced, un-verifiable content as expeditiously as possible. - Donald Albury 19:34, 26 January 2023 (UTC)[reply]
  • Oppose a blanket ban. I expect that in another five or six years, we will happily be allowing a Wiki-AI to both write and clean up most of our articles. BD2412 T 22:26, 26 January 2023 (UTC)[reply]
    And AIdmins to deal with AIsox ~ Selfstudier (talk) 22:53, 26 January 2023 (UTC)[reply]
    While all of this wouldn't surprise me given the other aspects of this dystopia we call Earth, it would simply confirm that we live in the Golgafrinchan Ark B world. Andre🚐 22:55, 26 January 2023 (UTC)[reply]
    I suspect it will become necessary to rely on such bots, as our current model is creaking under the current load. In just the last three or four days I've discovered a couple articles that I started that are in need of serious cleanup and repair, particularly on sourcing (linkrot and other issues) and updating. Donald Albury 22:58, 26 January 2023 (UTC)[reply]
    Oh don't get me wrong, I would love to see AI bots to improve references, and do other automated tasks under the bot policy. But the AI should not be trusted for facts or interpretations. And I do fear the ChatGPT-ization of the language. There is good and bad writing on Wikipedia, but at least it was written by and for human beings. Andre🚐 23:02, 26 January 2023 (UTC)[reply]
  • Oppose a blanket ban. I agree with King of ♥ ♦ ♣ ♠'s comment above. If an editor adds unsourced AI-generated content, that content should get treated the same way as non-AI generated unsourced content. WP:V and other existing Wikipedia policies already cover that. Some1 (talk) 00:05, 27 January 2023 (UTC)[reply]
    • But, a great deal of unsourced content doesn't get treated. The amount of unsourced content on Wikipedia is vast. If you don't believe me, pick a scholastic subject like Roman Empire, click on "What links here", and open a bunch of the links in tabs and start inspecting. What about random article? Too many stubs, and it's boring. With "What links here", you can get longer articles on average to view. The striking thing is the lack of "citation needed" tags - they are spaced few and far between. They can be found on roughly 1% of all pages, while unsourced content can be found on a much higher percentage.

      Another thing to try is go to Template:Unreferenced, and click on "What links here". The first page I clicked on was Tank destroyer. The tag is dated March 2009.

      The point is, you make it sound like all unsourced content gets quickly removed. That's not the case for a huge amount of content. It can sit there for years. LLM-generated content can be generated in great quantities fast, and therefore has the potential to accumulate more quickly than content composed by humans. Is it wise to let it sit there until a human comes along to remove it? In terms of a quantity competition between humans and computers, computers will win. This will take auto-removal to keep up. It would be best to start building those tools now. See Fram's post, above for an even better rationale.   — The Transhumanist   06:49, 28 January 2023 (UTC)[reply]
Holly and Judy, Wikipedians from the Glasgow branch, prepare for the AI wars.
  • Comment: There's no way to enforce a ban, and at any rate Wikipedia would do well to have more automation. Both MW and en.Wiki are heavily dependent on manual labor, with 3rd-party "bots" doing some of the more menial tasks. Compare how one would compose a document on a modern word processor vs. how editors do it here: no Wikitext, no copying-and-filling-out-templates (and no separate commits just for AnomieBOT to date them), no broken pages because of parsing errors, no dragging someone to ANI/AE for a T-ban violation (because there's such a thing as access control), no separate citations of the same source in five different formats (because there's such a thing as reference management); and you can actually comment on a specific paragraph without looking for a diff number, opening a "discussion" (which is in fact just another near-meaningless bit of Wikitext), signing it and hoping that no edit conflict arises because someone changed something 5,000 words up the page. We need to get rid of the concept of a WP:GNOME before we can even consider how to deal with a language model that can generate an entire article in a fraction of a second. François Robere (talk) 13:31, 27 January 2023 (UTC)[reply]
  • Oppose. I get that this raises some interesting questions in the abstract, but come on guys, we have enough problems with instruction creep already, we don't need to start writing policies in anticipation of the hypothetical abuse of Silicon Valley's latest fad. – Joe (talk) 14:18, 27 January 2023 (UTC)[reply]
  • Support blanket ban because fuck LLMs and fuck the corrupt, unethical industry that created them. XOR'easter (talk) 17:32, 27 January 2023 (UTC)[reply]
  • Oppose blanket ban blanket ban and suggest we develop Wikipedia:Large language models (perhaps under the name Wikipedia:Computer-assisted text generation suggested by Michael_D._Turnbull). I don't think that going forward we can ban AI-generated text writ large, first and foremost for the very simple example that many people currently editing wikipedia use text edit widgets that already incorporate something of this in the form of spell check, autocorrection and autocomplete, and these kind of tools will continue to blur the line between AI, language models, and human-generated text. Going forward it would be practically Neo-Luddism to eschew all AI. I don't like the use of ChatGPT right now today to generate text, I don't like it at all, but neither can I bury my head in the sand and whistle Dixie, pretending it doesn't exist and won't grow in importance. We should meet this head on rather than pretend we can completely ban AI-assisted or AI-generated text. —DIYeditor (talk) 23:27, 27 January 2023 (UTC)[reply]
  • Oppose What seems to scare people about ChatGPT is that it writes better than most Wikipedia editors. Competition is healthy and so should not be subject to restrictive practices. See also Luddism. Andrew🐉(talk) 22:32, 28 January 2023 (UTC)[reply]
    Competition is good when it's robot vs real people? You'd be happy if most of the encyclopaedia was written by AI? And most of the real people just gave up? Doug Weller talk 13:05, 29 January 2023 (UTC)[reply]
    I think that most people are at least open to the idea of using LLMs like ChatGPT for good and practical purposes, such as summarizing dozens of paragraph from a reliable source, or making a WP:Earwig-like bot that detect source-text integrity issues, or detecting possible hoaxes/context-dependent vandalism in Special:RecentChanges. I'm sure that when these LLM-based tools come out, people will use them just as much as mw:ORES and User:Cluebot NG today. The problem as of 2023 is that these tools does not exist yet and as in the current form, LLMs is an extremely powerful tool for bad actors yet disadvantaged good-faith Wikipedia editors. I feel that this situation between LLMs and Wikipedia right now a bit like Wikipedia and the academia in the early 2000s, when Wikipedia is full of uncited info and its reliability is really shaky to say the least (see also https://www.nostalgia.wikipedia.org). Maybe this will change in the future when whoever makes a LLM model that's aligned to our values and policies, but in my opinion for now a blanket ban is necessary to prevent mass vandalism while we are trying to process the situation. CactiStaccingCrane 13:24, 29 January 2023 (UTC)[reply]
@CactiStaccingCrane: How would a ban of LLMs prevent vandalism exactly? Vandals would simply ignore the ban, while many good actors would obey the ban. You would in effect be preventing good actors from using the tool, and not bad actors. The only way to deal with vandals who ignore a ban is directly – that is, identify their text and remove it, and block the vandals. But you can do that anyways. So, wouldn't it be best to identify and remove LLM vandalism while allowing good use of LLM-generated text? (See the starting edit and talk page for the chatbot-generated article Artwork title). So, I'm confused as to how you believe a blanket ban would help. Let me repeat my initial question, along with a follow-up question: How would a ban of LLMs prevent vandalism exactly? And why would banning good actors using LLM (like Pharos and JPxG) be necessary? I look forward to your replies to these 2 questions. Sincerely,    — The Transhumanist   21:46, 29 January 2023 (UTC)[reply]
Support blanket ban. Google and Bing actually considers auto-generated content as spam, which is bad for SEO. If anything, copying from GPT and pasting into an article is no different from copying from a free source and pasting into Wikipedia. The text generated is not your own, but is licensed to you forever, and, since this is a machine we are talking about and not a human, there is a probability that specific text output will be the same for multiple users for the same input.
I believe GPT is best used as an inspiration for further research, but in no ways is it actually any more useful than a machine translated article. Aasim - Herrscher of Wikis ❄️ 14:28, 3 February 2023 (UTC)[reply]
Oppose a blanket ban. I would prefer that editors use GPT to assist in writing, but to avoid plagiarism, tag their edits as "GPT-assisted". I would additionally suggest that GPT-assisted edits be all subject to pending changes, except if a user reaches Extended confirmed status. I personally fear that, for those who are familiar with American history, ChatGPT could be the prohibition era equivalent for Wikipedia. InvadingInvader (userpage, talk) 07:47, 6 February 2023 (UTC)[reply]
Pending changes can only be applied to articles. There is no way (that I am aware of) of having pending changes apply to only some edits to an article. Donald Albury 15:43, 6 February 2023 (UTC)[reply]

Chatbot This idea probably goes to the heart of it

How about this? (just a draft to be tweaked)

An editor is responsible for every part of every edit that they make. The norm is that they should have their own rationale for everything that they added including every word, phrase and sentence. For additions to articles, they should make a specific effort to make sure that the edit is appropriate with respect to the current article. For example, with respect to the structure of the article and avoiding duplication. They should also make a reasonable specific effort to assure that each potion of their addition is verifiable or verified in accordance with WP:Verifiability. It is unlikely that these requirements and expectations would be met with AI generated content (or any large amount of text that has been copied from elsewhere and pasted in, copyvio issues notwithstanding)
Wikipedia relies on volunteer efforts to review additions. This often requires time consuming reviews of individual words, phrases and sentences; obtaining this degree of effort relies on the understanding that the editor who put the material in has made a similar effort to develop that phrase or sentence. Expecting editors to give this review to large amounts of material which were generated by AI in a few seconds would cause a substantial loss of this effort.
Accordingly, this clarifies that removal / reversion of en masse additions of material suspected of being AI generated is considered to be an appropriate practice. An editor who seeks to restore the material is expected to break it into small portions, each with individual explanatory edit summaries. If such a removal results in deletion of the entire contents of the article, it then becomes a candidate for speedy deletion.

Sincerely, North8000 (talk) 21:07, 26 January 2023 (UTC)[reply]

I don't favour making machine-generated text a special case. Given that editors are already responsible for verifying every aspect of their edits, any clarifications should be equally applicable to all cases, such as human ghostwriting teams. isaacl (talk) 21:48, 26 January 2023 (UTC)[reply]
I like that a lot. — rsjaffe 🗣️ 05:06, 27 January 2023 (UTC)[reply]
I have integrated the above three paragraphs into Wikipedia:Large language models, see Wikipedia:Large language models#Specific guidelines and Wikipedia:Large language models#Summary removal of larger LLM-generated additions of article prose. —Alalch E. 10:33, 27 January 2023 (UTC)[reply]
cf. Wikipedia:Village pump (policy)/Archive 176#RFC: change "verifiable" to "verified". Also: do not make assumptions about the capabilities of AI. There are models at work that integrate references, and you should assume that at some point they'd be able to compose texts that are comparable to any Wikipedian's. Ergo, policy should focus on what we're looking for, not who or what composed it. François Robere (talk) 14:12, 27 January 2023 (UTC)[reply]
We can change the policy if something is developed that is reliable enough. Until then, blanket rejection is appropriate. — rsjaffe 🗣️ 15:46, 27 January 2023 (UTC)[reply]
It's important to see the conduct side too because LLM misuse forms a pattern of disruptive editing. It starts with one person's idea that a specific thing can be accomplished on Wikipedia in this way, proceeds with the intent to implement this idea without caring to understand and account for what Wikipedia's requirements are, and ends with an undesirable action which may be repeated if not addressed. —Alalch E. 15:55, 27 January 2023 (UTC)[reply]

If chatbots are banned, would the article Artwork title have to be deleted?

Artwork title was created by chatbot, and heavily edited by a human since. If chatbots (and LLMs) become banned, how would it apply to pre-existing chatbot-generated articles?    — The Transhumanist   22:09, 29 January 2023 (UTC)[reply]

I don't think it would have to be deleted (or restarted) unless it was somehow found to be a copyvio. Maybe it would be peer reviewed instead. Roll 3d6 (talk) 08:50, 5 February 2023 (UTC)[reply]

More specific proposal: Blanket ban on LLM content on Talk page discussions

Regardless of the community's decision on LLM-generated content in articles, which is the focus of much of the discussion above, the ability of editors to flood talk pages with artificially generated content arguing for a particular position on talk seems to have no redeeming value, and represents a new form of Wikipedia:Sockpuppetry. I propose a blanket ban on such writing, with especially strong guardrails for RfC's and AfD's. (Alternatively, I would be open to a phrasing that allowed LLM's to summarize the state of a debate, or be used to generate sample content for discussion, as in the conversation above, but not used to make arguments. That just seems harder to phrase clearly.)Carwil (talk) 20:22, 27 January 2023 (UTC)[reply]

That's a really weird way to use them. Also, why would it matter? Even if an editor used an LLM to generate a better phrasing of the argument they want, it's still their account putting forth the argument. And the argument is either valid or not in regards to others involved in the discussion. Why is this a problem exactly? Do you have any examples to better clarify this sort of usage? SilverserenC 23:14, 27 January 2023 (UTC)[reply]
Here's an article that should help clarify the relevant danger: Researchers demonstrate how attackers can use the GPT-3 natural language model to launch more effective, harder-to-detect phishing and business email compromise campaigns.. If they can use it to write convincing email scams, Wikipedia talk pages should be a breeze. Here's a quote from the article: ""The generation of versatile natural-language text from a small amount of input will inevitably interest criminals, especially cybercriminals — if it hasn’t already. Likewise, anyone who uses the web to spread scams, fake news or misinformation in general may have an interest in a tool that creates credible, possibly even compelling, text at super-human speeds." If that doesn't convince you, I don't know what will.    — The Transhumanist   10:16, 28 January 2023 (UTC)[reply]
Has this ever actually happened? – Joe (talk) 05:29, 28 January 2023 (UTC)[reply]
Maybe a more relevant question is "Will we be ready for it when it does?"    — The Transhumanist   10:16, 28 January 2023 (UTC)[reply]
As Transhumanist is suggesting, my concern is rapid creation of multiple arguments that either tilt a discussion or waste the time of sincere contributors. Users should be warned that they can't substitute mass-produced arguments for their own judgment inside the encyclopedia.--Carwil (talk) 19:26, 28 January 2023 (UTC)[reply]
  • Support - Weird? Compared to the many examples of what Chat-GPT can do posted on social media, writing talk page or forum posts on Wikipedia seems comparatively bland and simple. Why would using an LLM on a talk page or discussion forum matter? Because, it is faster than a human. With it, a user could participate in more discussions in less time. But, the big concern here is using it on multiple accounts with different writing styles to stack votes on issues with little chance of being discovered as the same person. That's sockpuppetry elevated to a higher level. Therefore, banning chatbots from being used to compose talk page or forum posts is quite reasonable.    — The Transhumanist   05:42, 28 January 2023 (UTC)[reply]
  • Support On a risk/benefit analysis, the potential benefit to allowing this is so small that any risk (as described above) is unacceptable. — rsjaffe 🗣️ 05:50, 28 January 2023 (UTC)[reply]
  • Reluctant Oppose As my comments above, it is difficult to draw a line on which language models are allowed and which are not. Clearly people are allowed to use autocompletion and predictive text and such. Are they limited in what varieties and designs of autocompletion they use? I think this requires further discussion and hopefully input from experts. —DIYeditor (talk) 06:42, 28 January 2023 (UTC)[reply]
  • Oppose a blanket ban as premature, but support at least some restrictions. While I would certainly be first in line to support a blanket ban if an actual problem arose, I am not sure that we know exactly what shape this will take enough to come up with an intelligent solution. For example, earlier in this very discussion, we were posting LLMs' output in order to judge their capabilities, so any prohibition would need to take exceptions like this into account. That said, I do support some more specific language for WP:LLM about it being very bad to use them undisclosed to argue your case in discussions. For example, if I were a shady dude, I could flood the zone with shit right here on VPP by typing out massive walls of text replying to every single person who disagreed with me, without regard for whether my arguments were sound or even correct, and even if I represented a minority view it would probably irritate and discourage my interlocutors until they stopped commenting (thus bringing me closer to a majority). Similarly, at the blood-soaked fields of AfD I could trivially write out a three-paragraph !vote on all 50 of the day's nominations (whether I was a partisan for keeping, a zealot for deleting, or a fanatic for any sort of POV). jp×g 19:45, 28 January 2023 (UTC)[reply]
  • Oppose a blanket ban. As JPxG, I'm open to some restrictions, but I can see LLMs as a potentially useful tool for people who want to offer their viewpoint in discussion but lack the fluency or time of some other editors. (A bit more acerbically, our discussions already tend to be influenced by editors who are functionally LLMs: good prose stylists, possessed of enormous sitzfleisch, and not well grounded in factual specifics. If LLMs force us to review WP:BLUDGEON and our methods of dialectic/achieving consensus, there's no reason to grant people like that the privilege of immunity.) Choess (talk) 20:44, 28 January 2023 (UTC)[reply]
  • Oppose per WP:CREEP. As a recent example, I included some ChatGPT output in a recent discussion at ITN. Not seeing the problem. Andrew🐉(talk) 22:23, 28 January 2023 (UTC)[reply]
I think your example misses the spirit of my proposal, which is un attributed LLM output substituting for our own reasoning and arguments on Talk. Happy to modify accordingly. --Carwil (talk) 13:25, 29 January 2023 (UTC)[reply]
  • Oppose a blanket ban on talk pages specifically (I would support one in article space). The largest problem by a huge margin with using ChatGPT on Wikipedia is adding factually incorrect information to articles. This is less of a concern with talk pages, which the general public doesn't usually read and which are not primarily meant to impart facts or be reference material. We already have a ban on sockpuppeting and an expectation that decision-makers will ignore bad arguments, which cover the negative use cases mentioned above. Also, as mentioned above, there is not an enormous difference between ChatGPT and predictive text or services like Grammarly, and the line between those is going to become blurrier by the day, faster than policy can keep up with. Gnomingstuff (talk) 01:20, 3 February 2023 (UTC)[reply]
  • Oppose as making up rules when no problem has been demonstrated. If it does happen, we can warn, revert, and block the offending material and offender for wasting our time. Graeme Bartlett (talk) 12:10, 3 February 2023 (UTC)[reply]

The discussion has diverged

Further issues on this topic (chatbot-generated content) are being discussed at Wikipedia talk:Large language models, the talk page for the policy draft on this subject.    — The Transhumanist   05:15, 30 January 2023 (UTC)[reply]

The discussion has been re-opened

Allow registered editors to use vpn (open proxies)

Currently, WP:PROXY states about itself, "This policy is known to cause difficulty for some editors who may need to use open proxies". I have experienced said difficulty and whenever I try to use my vpn I get the Wikipedia notice that I am not able to edit because of it. The rationale of the policy states, " open proxies are often used abusively. MediaWiki, the wiki software that powers Wikipedia, depends on IP addresses for administrator intervention against abuse, especially by unregistered users." Why not let registered editors use vpn (open proxies)? When I use an online website with interaction with other users, oftentimes I can block any given user, I don't need their ip. I don't see why Wikipedia cannot do the same by just blocking the account without resorting to the ip.

The current policy and technical actions of blocking the use of open proxies by registered users seems to be unreasonable. Placing bureacratic hurdles to be able to use one seems to be also unneeded and unreasonable when it comes to said registered editors.

Many people who want to contribute to Wikipedia are probably just enthusiastic about editing here but they may not have much idea in what kind of serious problems that can even destroy their life they can get in by the simple act of editing.[1] [2]

As a balance between vandalism by anonymous users and the safety of editors, Wikipedia should allow unrestricted vpn (open proxies) use by registered editors. Thinker78 (talk) 17:11, 9 January 2023 (UTC)[reply]

@Thinker78: This is pretty much a nonstarter. The IP is needed to help prevent block evasion via WP:SOCKPUPPETRY. If you really need edit behind a proxy, then simply just request WP:IPBE.
It's also not true that other websites don't block VPNs. Netflix routinely does this as well (albeit for very different reasons). –MJLTalk 19:52, 9 January 2023 (UTC)[reply]
I never said that "other websites don't block VPNs". Besides, what's the percentage of registered editors who have been blocked? What's the percentage of those blocked editors who could cause really harm to the project for suckpuppetring as opposed to any ip user or new account? Also, for any issues with a new sockpuppet account, pages can be protected.
According to WP:IPBE there are only 806 editors with the block exemption out of the millions of editors in Wikipedia. That's not a very successful statistic of the program. The balance to be made is between sockpuppetry and the safety of editors. Thinker78 (talk) 22:09, 9 January 2023 (UTC)[reply]
The people who are blocked by this are disproportionately from developing countries. See m:Talk:No open proxies/Unfair blocking for some examples.
This is going to become a bigger problem. Blocking everyone who uses Apple's iCloud Private Relay is going to cut into the English Wikipedia's core editor base. We're asking people to choose between disabling privacy features on all websites, or not being able to edit. Google Chrome, which is the most popular web browser among editors, is likely to ship something similar in the next year or two. MediaWiki (the software that we use) may have to stop focusing on IP addresses and move to another system, like a Device fingerprint.
I know that there's been talk among the devs and product managers about this problem recently (also, off and on for at least ten years). The one thing that any registered editor could do to help in the short term is to turn on the "IP Info" item in Special:Preferences#mw-prefsection-betafeatures. I've found this answers most common questions (e.g., what part of the world is this IP editing from?), and it's really handy on the history page. Please try it out, and provide your feedback to the team, so they can get this initial project wrapped up. Whatamidoing (WMF) (talk) 03:38, 10 January 2023 (UTC)[reply]
I live in Guatemala and many times when I try to edit from my cell phone using data and not wifi I find that I am blocked. For some reason the ip address assigned to my phone by my ISP is not of the liking of the blocking code or the dev who placed the restrictions. Basically people are blocked from editing from cell phones only because they live in Guatemala. Thinker78 (talk) 15:49, 10 January 2023 (UTC)[reply]
We have the same problem in the US. If you use T-Mobile (the second biggest mobile phone service in the US), then you can't edit from your phone. Admins have blocked all the IPs. Whatamidoing (WMF) (talk) 22:42, 11 January 2023 (UTC)[reply]
With respect it's nowhere near the same problem. T-Mobile is just about never hard blocked. You just need to log in (and if you think that needs softening then you'd be wrong). -- zzuuzz (talk) 22:51, 11 January 2023 (UTC)[reply]
That sounds like something they should take up with their VPN client provider. Many clients allow you to whitelist destinations. — xaosflux Talk 15:56, 10 January 2023 (UTC)[reply]
That might be possible if we could safely assume that they have control of the VPN (and, e.g., not the VPN that 30% of internet users say they are required to use for work), and if we assume that someone who wants to add a sentence or correct an error in an article has enough skill to know how to do that. Just from those two groups, I'd guess that less than half of people are able to do this in practice. Whatamidoing (WMF) (talk) 22:37, 11 January 2023 (UTC)[reply]
Whereas other sites require some form of identity verification, whether through email, phone, payment, social media or other checks and metrics, Wikipedia requires none of that. The requirements for an account here are basically non-existent. For this reason, users who are merely registered or unregistered are usually more similar than you might think. -- zzuuzz (talk) 20:17, 9 January 2023 (UTC)[reply]
Some of that identity verification is meaningless though. Most require email verification, but that email could be a throwaway email or to a disposable email service, so not really good identity verification. For better or worse, Wikipedia's VPN/proxy blocks are definitely stronger/better than any other site I've seen, and are a stronger form of identity verification. ProcrastinatingReader (talk) 15:18, 29 January 2023 (UTC)[reply]
This three-part series on proxy blocks from Vermont on Diff may be relevant: first part, second part, third part. 🐶 EpicPupper (he/him | talk) 05:25, 10 January 2023 (UTC)[reply]
You are linking to the same article for the second and third parts. 2603:8000:FC00:800:1055:C75E:A3A:3DE (talk) 17:39, 10 January 2023 (UTC)[reply]
Good catch, thanks. 🐶 EpicPupper (he/him | talk) 01:51, 11 January 2023 (UTC)[reply]
Those are great blog posts. BTW, the blog's open to volunteers. If you know something about this problem (or another subject) and want to write it up, please contact User:RAdimer-WMF. Whatamidoing (WMF) (talk) 22:44, 11 January 2023 (UTC)[reply]
I have been caught out on this a few times myself due to the iCloud relay before giving up on mobile or iPad editing. I was initially thinking that the idea was good but once reading the replies made me realise the issue was more complex. My immediate thought was that, perhaps, it could be Extended Confirmed required to use an open proxy (or carrier grade NAT or similar). If an extended confirmed user gets blocked then their user created for block evasion would still need to reach EC before being able to bypass anything. Gusfriend (talk) 02:09, 11 January 2023 (UTC)[reply]
Unfortunately we already have disruptive editors who game extended confirmed so that they can edit disruptively on articles that are EC protected. Making it so that upon granting of EC the editor is also given the equivalent of IPBE would make it harder for CheckUsers to identify repeat offenders and block any sleepers they may have pre-prepared. Sideswipe9th (talk) 01:32, 15 January 2023 (UTC)[reply]
How do they game extended confirmed? Besides, the issue cannot just be thought about restricting editors but there has to be a meaningful balance with the safety and privacy of editors online. Thinker78 (talk) 03:22, 15 January 2023 (UTC)[reply]
Same way disruptive editors game autoconfirmed. I'll not say more than that per WP:BEANS, just that the difference between the two types of gaming is one of volume and time. If you edit any of the more controversial topic areas though, I'm fairly positive you will have seen this behaviour.
I agree that there is a balance to be struck between safety and privacy, but I do not agree that granting all extended confirmed users meets that balance. IPBE is closer to balancing those needs, though it is also not without its problems. Sideswipe9th (talk) 03:31, 15 January 2023 (UTC)[reply]
"it could be Extended Confirmed required to use an open proxy (or carrier grade NAT or similar)" – I agree strongly with this. It's time Wikimedia left the 2000s and entered the 2020s on security, but we don't want randos being able to hide their IPs from us.  — SMcCandlish ¢ 😼  23:46, 13 January 2023 (UTC)[reply]
We already have WP:IPBE though. If you are effected, then apply. If lots of people get denied this, then we need to rethink, but if we're just complaining about it without applying it's a bit of a non-starter. Lee Vilenski (talkcontribs) 01:10, 15 January 2023 (UTC)[reply]
As I mentioned previously, "According to WP:IPBE there are only 806 editors with the block exemption out of the millions of editors in Wikipedia." Thinker78 (talk) 03:27, 15 January 2023 (UTC)[reply]
That's not a relevant statistic though as the vast majority of those editors have no need for IBPE (or had no need at the time they edited). We have a current system (IPBE) and so before replacing it with something else we need to determine whether it is fit for purpose and, if it isn't, whether it needs modifying or replacing. Some questions that should help determine whether it is fit for purpose are:
  1. How many people have a need for IBPE (or a replacement)?
  2. Of those, what proportion are good faith editors? (because at least some bad faith editors would definitely benefit in ways we do not want)
  3. How many of those good-faith editors know to ask for IBPE?
  4. Of those good-faith editors that ask, how many are approved?
I don't pay a lot of attention to the IPBE discussions that happen on the Functionaries mailing list, but from what I remember the answer to 4 is that it is at least a majority and probably a significant one. Thryduulf (talk) 12:45, 15 January 2023 (UTC)[reply]
That statistic doesn't include admin, who have the right by default. The question is how many people are being denied access via a IPBE request? Lee Vilenski (talkcontribs) 13:47, 15 January 2023 (UTC)[reply]
@Lee Vilenski admnins have "half" of IPBE; they can bypass blocks but can't use tor. — xaosflux Talk 17:03, 15 January 2023 (UTC)[reply]
What's your standard to weigh whether someone needs IBPE? Thinker78 (talk) 17:47, 15 January 2023 (UTC)[reply]
My personal standard is that IPBE is needed only when someone is unable to contribute (safely) without it and will use it only to contribute in good faith. The second is obviously hard to judge objectively, especially for new users, but for existing contributors your contribution history is usually a fairly reliable guide. (Note this is not an official position of the functionaries or any other group). Thryduulf (talk) 15:36, 16 January 2023 (UTC)[reply]
How is the determination if someone is unable to contribute safely? Examples? Thinker78 (talk) 18:44, 16 January 2023 (UTC)[reply]
There is and can be no single criterion for this as it depends a lot on individual circumstances, but an editor in a first world country wanting to write about first world topics can do so safely without IPBE in pretty much every case (but exceptions may exist). In contrast an editor in Iran wanting to write about contemporary Iranian politics will plausibly need to use a proxy for their own safety. Thryduulf (talk) 20:46, 16 January 2023 (UTC)[reply]
I think that standard is unreasonable. For starters, security experts recommend using vpn when connecting to a network one doesn't trust (it could be public wifi, at home with relatives or roommates one doesn't trust, or a snooping ISP who wants to sell data).[3]
In addition, regarding first world countries, you probably didn't hear of the time when the government of France forced an editor of the French-language Wikipedia, and president of Wikimedia France, under threat of detention and arrest, into deleting an article about Pierre-sur-Haute military radio station.
This points out that even in first world countries, specially in the US, people are subject to violations of their human and legal rights regularly by authorities. Police regularly arrest people for legally protected things, like taking pictures in public places, for having a bumper sticker in the truck.[4] or for criticizing public officials[5]
Finally, some US law enforcement authorities are deeply corrupt and can even murder people for criminal organizations.[6] Thinker78 (talk) 23:03, 16 January 2023 (UTC)[reply]
[6]. 🐶 EpicPupper (he/him | talk) 23:46, 16 January 2023 (UTC)[reply]
@EpicPupper that's like the dilemma as to whether to have bodyguards or not. How would one know the bodyguards are not bought out or tracking one's every move for nefarious purposes or even to plan to takeover one's businesses? Some bodyguards are really great people, but others probably not so much.
The issue then becomes of choice. The normal advice is to use an open proxy if one thinks the connection is not safe. Some providers of vpn may be dishonest but others are not. Whether its use gets to be beneficial or not really depends on individual circumstances. Thinker78 (talk) 19:57, 17 January 2023 (UTC)[reply]
The normal advice is to use an open proxy if one thinks the connection is not safe. @Thinker78: VPNs don't provide security.

You mention bad acts by authorities as part of your justification; if you are concerned about the government arresting you, a VPN will not help. It will simply move the organization that a government needs to order to hand over evidence from a website to the VPN provider (see one highly-publicized example; the service is still in operation). In that case, I would encourage you to ask yourself whether you trust the Wikimedia Foundation or an arbitrary VPN provider of your choice more. 🐶 EpicPupper (he/him | talk) 03:15, 18 January 2023 (UTC)[reply]
In the US, the government needs a warrant to compel a business to hand over data. Although sometimes data is handing over by simple request, the point is that there needs to be an extra step rather than without having any privacy protection.
I don't quite see the connection presenting the Wikimedia Foundation vs an arbitrary VPN provider. What's your point? (I had made the following comment previously: "I know that Wikipedia has a secure connection but it has happened to me that the network transaction to connect to it sometimes is reset in the middle of it and the connection becomes unsecured for a couple of seconds, enough to reveal the contents of whatever I am doing at the time.") Thinker78 (talk) 19:05, 18 January 2023 (UTC)[reply]
My point is that your data can be handed over through a warrant, with VPN or without. The question is simply *who* is warranted, the WMF or a VPN provider. 🐶 EpicPupper (he/him | talk) 23:21, 18 January 2023 (UTC)[reply]
That would be a situation specific to the US. Because foreign governments probably would have a harder time to make Wikimedia submit user data. And I happen to edit from Guatemala. My point was simply that even in first world nations there is risk by the government or other organizations. My guess is that the WMF would make it harder for the US or local government to be able to get user data.
If the user doesn't use an open proxy there is still the issue of the ISP which is probably all too happy to hand over whatever data the government asks for, unlike most vpns or proxies. And as I pointed out, there are many other uses of open proxies that can protect the privacy and data integrity of editors. Thinker78 (talk) 03:15, 19 January 2023 (UTC)[reply]
the WMF would make it harder for the US or local government to be able to get user data. Great, then a VPN would compromise your privacy if they have looser rules than the WMF in terms of data disclosure.
many other uses of open proxies that can protect the privacy and data integrity of editors. please mention these uses.
the issue of the ISP which is probably all too happy to hand over whatever data the government asks for, unlike most vpns or proxies. please provide evidence that "most vpns" will not happily hand over user data. I've linked the HMA case, see also this example of a public disclosure policy. 🐶 EpicPupper (he/him | talk) 03:27, 19 January 2023 (UTC)[reply]
Uses of open proxy for non-government issues, for example, connecting to a network one doesn't trust or presents issues in a restaurant, cafe, airport, street, roommates, shady relatives, library, work, college, etc.
You asked about trust, well, I trust more certain open proxies than ISPS regarding illegitimate government requests. And I believe that open proxies because of the nature of their business are somewhat more trustworthy in general than ISPS. Not saying that some vpns can't be honeypots or shady. Thinker78 (talk) 03:43, 19 January 2023 (UTC)[reply]
I'll add my own personal criteria, which is slightly different to Thryduulf, and probably on the more liberal side of written policy. The user generally needs to be a) unreasonably affected by blocks, and b) not a sockpuppet. There's obviously some judgement which goes into that. I'm not hugely sympathetic to the first world problem of not being able to edit through Apple relays from your latest iPhone due to some false sense of security. Be reasonable: Just turn it off. It's not going to get you arrested, and it will help checkusers. Also, trust me, you don't want to be using the same IP addresses and user agent as some banned troll making death threats. I'm open to persuasion in individual cases though. For users in countries strongly affected by P2P blocks, or in countries with serious security issues, I'll often grant IPBE without even a request, subject to my second criterion. As for who is not a sockpuppet? it's probably one of those things you know it when you see it. I've seen many many extended-confirmed socks, some of whom request (and sometimes get) IPBE. There's no real threshold, short of several years and many thousands of edits, where I'd place any automatic grant. For one thing, we have a plague of accounts getting compromised, and having IP blocks can help when that happens. I find it interesting looking through all the IPBE grants made by all the admins above. I count maybe two in the last several years. There's probably several factors at play there. Lack of requests is probably a leading factor. Also, talking of statistics, the number quoted is just a snapshot. Many IPBE grants come and go as, like a lot users, they're usually temporary. -- zzuuzz (talk) 20:27, 16 January 2023 (UTC)[reply]
Zzuuzz, as per my previous reply, there are no guarantees that no one gets arrested or worse for what they contribute in Wikipedia in the US. Thinker78 (talk) 23:07, 16 January 2023 (UTC)[reply]
If the authorities are watching what you're doing through your account and HTTPS then you have bigger problems than your IP address. I'll wager that using Apple services won't help your situation either. Most of what you mention is unrelated to IP addresses. I'll just add that I have granted IPBE for these types of concerns. They're not common, and are rarely plausible. -- zzuuzz (talk) 23:39, 16 January 2023 (UTC)[reply]
It's more about people not in the radar of despotic or corrupt authorities not getting in their radar. And if they are in their radar, make it more difficult to get snooped on and not completely forget about cybersecurity. I know that Wikipedia has a secure connection but it has happened to me that the network transaction to connect to it sometimes is reset in the middle of it and the connection becomes unsecured for a couple of seconds, enough to reveal the contents of whatever I am doing at the time. Thinker78 (talk) 20:03, 17 January 2023 (UTC)[reply]
@Thinker78: That is not how HTTPS works; the connection will not "become unsecured for a couple of seconds" (Wikimedia sites use HSTS). Please familiarize yourself with the underlying technology before making similar claims.

It looks like you've brought up the "authorities" lens again. Please see my comment above.

If your concern is privacy through hiding your IP addresses, IP addresses are not private information and cannot be easily linked to personal details. The linked Google Public Policy Blog post is from 2008; the Web has changed even more since, with more address reuse by ISPs.

Ultimately, the movement is not well-equipped to deal with matters of personal preference or of "societal issues" with regards to governments. If individual editors want to edit with a VPN, they can ask any of the 498 active administrators to request IPBE. 🐶 EpicPupper (he/him | talk) 03:23, 18 January 2023 (UTC)[reply]
Thanks for your reply. I have to point out that instead of assuming that I am not familiar with how HTTPS works and discouraging me to make "claims", it would have been more helpful to direct me to Phabrikator or a relevant venue. Regards, Thinker78 (talk) 19:51, 18 January 2023 (UTC)[reply]
I do not see how Phabricator is relevant in this situation, hence I did not link to it. 🐶 EpicPupper (he/him | talk) 23:20, 18 January 2023 (UTC)[reply]
I'd also like to note that making broad and incorrect claims can incite worry amongst readers of this page; I did not caution against them without reason. 🐶 EpicPupper (he/him | talk) 03:28, 19 January 2023 (UTC)[reply]
There is no guarantees that no one gets arrested or worse for what they contribute anywhere, whatsoever. It's not just the US, all countries suffer from this problem. --RockstoneSend me a message! 04:24, 25 January 2023 (UTC)[reply]
@Thryduulf and @Zzuuzz, this story is for you:
Way back in the day, I talked to someone in what's now m:T&S about a problem he'd resolved. The WMF had received an order from a US court to hand over the IP addresses of everyone who had ever read(!) a particular Wikipedia article.
In the end, that information is automatically purged after no more than 90 days anyway, and the law enforcement agency was satisfied with having their over-broad order complied with more narrowly (i.e., only IP addresses that could be connected to the relevant local area), and, happily, it turned out that nobody in that area had read that particular article, so they ended up not having to disclose any IP addresses at all.
But think about that for a moment: The WMF received a lawful court order to disclose the IP addresses of every single person in the world who had read that article.
When you say "Just turn it off" or that most editors can edit "safely without IPBE in pretty much every case", I'd like you to remember that it's not always possible to even read safely. Most people's idea of safety and privacy doesn't include having law enforcement asking you to explain why you were reading an article related to a violent crime they're investigating. And while this story had a happy ending for our users' privacy, if we say "Just turn if off", we're asking them to turn off privacy tools for every website, not just ours. We intentionally purge our IP addresses automatically. Many websites make the opposite decision, and keep that information forever. IMO the ability to contribute to Wikipedia should not be constrained by your willingness to have your location tracked by advertisers or anyone else. Whatamidoing (WMF) (talk) 22:07, 17 January 2023 (UTC)[reply]
@Whatamidoing (WMF) readers do not require IPBE to read articles.

VPNs do not provide privacy through their obsufucation of IPs; IP addresses are not private information and cannot be easily linked to personal details. The linked Google Public Policy Blog post is from 2008; the Web has changed even more since, with more address reuse by ISPs. Also note that most advertisers already know your location regardless of if you use a VPN; billions of pieces of information have already been collected from users by the time they turn 18.

I see that you posted this message with your WMF account; is this, then, a statement in your capacity in the Community Relations team, and is it intended to represent the team's thoughts on this matter?

Thanks! 🐶 EpicPupper (he/him | talk) 03:31, 18 January 2023 (UTC)[reply]
Did anyone say that IPBE is necessary to read articles? I certainly didn't.
I tell this story to illustrate that we should not assume that IP addresses are ever safe or harmless things to disclose, or that any of us can easily judge when or whether an editor "needs" IPBE. Do I personally want to edit from an IP range that also contains a malicious actor? No. Do I want to exclude every current and potential editor whose ISP that uses CGNAT or who edits from an iPhone, and who hasn't been able to both discover the process for requesting IPBE and then convince an admin that they really, truly, desperately "need" IPBE? Also no.
As for your other claims, most IP addresses can no longer be easily linked to personal details, but this is not always true, even today. Whatamidoing (WMF) (talk) 03:52, 18 January 2023 (UTC)[reply]
  • I don't this this is a great idea as-is, but perhaps something. We could make some intermediary access, that only has (ipblock-exempt), and does not have (torunblocked, sfsblock-bypass) - but would need to decide on some bar for it... so it would need to be what does the community and admin who have to deal with this is a low enough bar? I don't think econf is high enough - maybe something like 1000edits and 180 days with autopromoteonce. If the bar is so high that it would be useless, then it is useless to bother though. — xaosflux Talk 17:09, 15 January 2023 (UTC)[reply]
    Maybe 90 days? Half a year seems like quite a high bar. 90 days would already be 3x XC and more than inconvenient for anyone trying to game the system. Terasail[✉️] 17:33, 15 January 2023 (UTC)[reply]
    According to Wikipedia:List of Wikipedians by number of edits, only 0.25% of editors are extended confirmed. That's 1 in 400. Now the questions are, what range of number of edits do you think is entry level bar for ipblock exempt? What's the global share of edits performed by said segment of editors? To have a rough measure of good faith edits, what's the percentage of said editors who have been blocked indefinitely (keep in mind that even User:Koavf was at one point blocked indefinitely)? Thinker78 (talk) 17:44, 15 January 2023 (UTC)[reply]
However this policy is definitely causing more difficulties for unregistered users than registered ones, especially those who needs to use proxies to bypass the censorship of Wikipedia. They will feel frustrating after trying to create an account but find his IP blocked. There are many ways to bypass the block, but WP:IPBE is only applicable for registered users and WP:UTRS is very inconvenient. They may even be unable to send emails to unblock-en@wikimedia.org as the address is blocked. The only way may be requesting to unblock on his talk page on Meta-wiki, but it is impossible for administrators to unblock the proxy so they may give advice to the user to let him try the IPBE, UTRS, etc. In conclusion, it is very hard for an unregistered user to contribute to WP at such situations. (I know these because I also have had such experience.)
Then I would like to share what I think is useful to directly connect to Wikipedia: TCPioneer (found on an instruction page on the Chinese WP) IntegerSequences (talk | contribs) 07:08, 19 January 2023 (UTC)[reply]

I think we definitely need to be realistic; one of the biggest problems Wikipedia has is its tendency to solipsism even when it's not at all helpful. We already recognise that editors behind the Great Firewall have a need for editing behind a VPN. As WhatAmIDoing points out, there are times where ultra-zealous law enforcement agencies have been casting wide nets already, and with the direction of politics in some places of the developed world… well, what if a red-state attorney general decided that editing Wikipedia articles about abortion or trans issues constituted promotion of the idea, and demanded editor details? What if Suella Braverman decided that the article English Channel migrant crossings (2018–present) (edit | talk | history | protect | delete | links | watch | logs | views) was "portraying people smugglers in a positive light"? We can't ask editors to make themselves feel unsafe just because we want to stick to a view of "open proxies" that's been obsolete for at least half a decade. That said, I understand the need for a balance; at the very least, IPBE should be shall-issue if someone in reasonable enough standing requests it. Sceptre (talk) 09:04, 23 January 2023 (UTC)[reply]

  • I think the blocks are irritating, and will be increasingly so. I use iCloud's proxy, for example, and hence sometimes encounter these blocks, at least on a mobile device. I used to disable the proxy temporarily + restart my browser, these days I just wouldn't bother make an edit on a mobile device. I suspect this feeling is much stronger for driveby editors who might see an error, wish to fix it, but don't really want to spend minutes changing system settings to do it, so will abandon. The CU use-case makes it difficult to just abandon these blocks I suppose, but the problem stands. ProcrastinatingReader (talk) 15:24, 29 January 2023 (UTC)[reply]

There is an RfC on whether Vector legacy should be restored as the default skin on the English Wikipedia. If you would like to participate in the discussion, you are invited to add your comments at Wikipedia:Requests for comment/Rollback of Vector 2022. Thank you. InfiniteNexus (talk) 00:49, 20 January 2023 (UTC)[reply]

Should news sources without attribution to specific authors be deprecated?

There is a lot of discussion currently about the risk of generated content being entered into Wikipedia directly, but another concern is third-party generated content being used indirectly as a source. Just last week there was some level of reporting regarding CNET's now-retracted decision to publish articles on their site that were written by an "Article Writing AI", articles which, despite ostensibly being screened by an editor, were found to have significant factual errors. CNET attributed these articles to CNET Money Staff, and then, after some criticism, just CNET Money. Now, because of the negative attention, they've "paused" the rollout.

https://futurism.com/cnet-ai-errors

https://www.theverge.com/2023/1/20/23564311/cnet-pausing-ai-articles-bot-red-ventures

The thing is, CNET is still considered a reputable publication, which is why it was notable that they did this and is also why they publicly stopped. However, there are certainly many other smaller sites that could implement technologies like this (and maybe already have) in order to appeal to SEO and search result rankings, and as these sources suggest, it will be very easy for factual errors to slip by well-meaning editors who are not used to manually fact-checking every single factual statement in a given text.

Given this issue, I think we may want to consider deprecating sources from news/newsblog articles that are not attributable to a specific author, or at least doing so for sites that have been confirmed to use AI generated content. Tpdwkouaa (talk) 18:38, 23 January 2023 (UTC)[reply]

Or, we could wait and see if sites lose their reputation for reliablity because they have been found to be using AI generated content that contained unacceptable levels of false information. I am very leary about deprecating sources because of what they might do, rather than because they have demonstratably published false content. Donald Albury 19:30, 23 January 2023 (UTC)[reply]
I have to agree with Donald… WAY too soon… we can react to what actually happens, and should not “pre-act” due to fear of what might happen. Blueboar (talk) 19:44, 23 January 2023 (UTC)[reply]
If a source becomes unreliable for this practice, then it's not suitable for use. That's the time to depreciate. We shouldn't do so on the idea that it might not be what we think. Lee Vilenski (talkcontribs) 19:47, 23 January 2023 (UTC)[reply]
Much too broad a solution to this narrow issue. Newswire services generally do not credit a specific author - this change would deprecate a very large number of highly reliable sources. I would support deprecating CNET even if they've stopped their article-writing bot, since clearly their editorial oversight is quite poor and that has implications even for articles written by humans. Ivanvector (Talk/Edits) 19:52, 23 January 2023 (UTC)[reply]
Overreaction to a minor issue that will complicate things immeasurably. Many news sources simply have their author set as "News desk" or something generic like that. Similarly, print newspapers often don't directly name an author for the majority of clippings. Should instead be handled on a case-by-case basis. Curbon7 (talk) 19:53, 23 January 2023 (UTC)[reply]
I've long stated that the biggest threat is that Wikipedians seems to have no idea what to do now that 2rd party vetting and verification is becoming ever less reliable. In the mean time, there are some great reporters who seem to have switched to self publishing, but we don't recognize that work either. We need to begin putting thought into replacing that entire system, probably by creating a new system to allow individuals to verify sources, news reporting and news reporters and probably keep track of some sort of ranking and reputation score or something like that. The fact that we keep being stuck in old conventions and refuse to even think about what first hand and oral knowledge means is concerning for our future, because its going to get pretty thin in sourcing when half the news is coming out of tik tok reporters. —TheDJ (talkcontribs) 22:33, 23 January 2023 (UTC)[reply]
Reliability is in general an issue, I don't like "Staff" or "News Desk" either but the suggestion is overkill. Sure, we should be keeping an eye on this and if needs be, case by case, addressing it. Selfstudier (talk) 22:40, 23 January 2023 (UTC)[reply]
  • There are dual, growing problems: the old Gell-Mann Amnesia effect (worsening, due to persistent cost-cutting and monetary pressures), and the new (partial) downfall of "gatekeepers" (i.e., formal publishing of any kind).
Here's how we could start to tackle that: WP:RSN is too rigid a mechanism. Most WP:RSN RFCs judge a source's quality based on a handful of articles, which is a minuscule amount of the source's overall content; they also largely rely on vague heuristics, like how the source's website looks, whether it's in text or video format, whether the authors have formalcredentials, etc). I can come across 10 missstatements of fact by The XYZ Post over a few months, but I'm not keeping notes, so those falsities won't be discussed in any RFC. And I can come across outstanding reporting by a self-published source or blog, but if I miss that RFC, it'll never be discussed.
We should create a new page, with subpages for each source (broadly defined; could even include YouTubers). When a source published an incorrect statement of fact, we list that to their subpage; when they publish high-quality exclusive reporting (again, even including YouTube videos), we list that too. This would allow any WP:RSN RFC to judge a source far more holistically and more comprehensively than is done currently. (note: this is only a very early idea, not a proposal) DFlhb (talk) 14:03, 25 January 2023 (UTC)[reply]
  • If the problem is with sources generated by AI then the solution should be targeted at sources generated by AI, not at the other 99+% of news sources without attribution to specific authors. There may be other reasons for not accepting news sources, whether or not they have attribution to specific authors, but this proposal is not about those reasons. Phil Bridger (talk) 23:08, 23 January 2023 (UTC)[reply]
    And there are tools that, at least for now, can detect AI-generated text, so questionable articles can be tested to see if an AI is probably involved. See https://openai-openai-detector.hf.space/, for example. — rsjaffe 🗣️ 23:19, 23 January 2023 (UTC)[reply]
  • Just to be clear, most old newspapers (which are widely used as sources for old biographies) identify no authors. I assume this question is directed at modern or future sources. BD2412 T 23:40, 23 January 2023 (UTC)[reply]
  • This is very common with syndicated news articles written by news agencies (e.g. the Associated Press, Agence France Press, Reuters), which are actually probably the most reliable sources in news. See e.g. a syndicated article by the AP published in the SCMP, or a syndicated article by AFP in France24, neither of which have a named author in the byline. So this proposal is a complete nonstarter. Endwise (talk) 00:21, 24 January 2023 (UTC)[reply]
    The number of times I have found factual, contextual and positional (bias) errors with stories (signed or unsigned) by the news agencies listed above is too big to count. Sometimes they disseminate worthwhile information. Sometimes they are just worthless. Since there is no mechanism to discern in advance which of those conditions apply for any specific story, they can never be called "reliable". That is a blanket statement that is just not true. On the issue raised by the OP I agree that the proposal is too broad. 65.88.88.59 (talk) 20:55, 24 January 2023 (UTC)[reply]
    The fact that a source sometimes makes mistakes does not make it unreliable. As it stands, mainstream news agencies are considered generally reliable by the Wikipedia community: WP:RSP lists Agence-France Press, Associated Press, and Reuters as generally reliable. If you think that consensus is wrong, you need to start a new discussion about that, preferably at WP:RSN, but you're going to need strong evidence to convince people.
    Insofar as mainstream news agencies are currently considered reliable, a new policy which makes them unreliable as an accidental side-effect is clearly a non-starter. Caeciliusinhorto-public (talk) 09:59, 26 January 2023 (UTC)[reply]
    Agreed with Endwise, and I'd have used the exact same words ('complete nonstarter'). Our current mechanisms are beyond sufficient: if a source is found to use AI bots, WP:RSN will deal with it accordingly, including by determining which properties are affected (whether a whole outlet, or just a sub-brand). DFlhb (talk) 14:16, 25 January 2023 (UTC)[reply]
    No, we should not deprecate news stories with no author. Many news agency stories are team efforts with no byline. Deprecating sources with no author would also wipe out most sources used in our coverage of the Russo-Ukranian war. I do agree with the comment above that many very reliable authors are publishing independently and are too easily dismissed as "blogs", but that is a different issue. Elinruby (talk) 23:14, 30 January 2023 (UTC)[reply]

Elevate WP:BLUDGEON to guideline

The essay WP:BLUDGEON seems to have wide support among the community and compliance with it is expected as part of standard user conduct. I was surprised that it's not already listed as a guideline. Despite this, I still often see users, sometimes even well-established users, replying to and arguing with a dozen comments in large discussions. I propose that we make it official and include it as a guideline. I believe it could be done by either of the following:

Thebiguglyalien (talk) 19:08, 24 January 2023 (UTC)[reply]

I don't think this would be a good idea. Much of the time people who oppose a position will accuse an editor advancing it of bludgeoning the process, but those supporting it will, on the same evidence, regard that editor as simply correcting the point made. Neither gets us any closer to a resolution. Phil Bridger (talk) 20:09, 24 January 2023 (UTC)[reply]
IMO not a good idea. Any persistence or failure-to-cave can be claimed to be bludgeoning. North8000 (talk) 20:21, 24 January 2023 (UTC)[reply]
I don't think that it would be pointful to make such a change because Wikipedia:The difference between policies, guidelines and essays. WhatamIdoing (talk) 22:51, 24 January 2023 (UTC)[reply]
This is the sort of thing that I think should be fixed. It's unfair to users if we say not to present essays as policies but we then use essays to justify sanctions when it's convenient. Thebiguglyalien (talk) 01:26, 25 January 2023 (UTC)[reply]
But, the purpose of having an essay written out is so that people don't have to type out the entire essay every time they wish to justify something that the essay explains well. Instead of saying, "I think this person's actions are disruptive because they are repeatedly commenting on the discussion, to the point where no one else can get a word in edgewise. Clearly, that kind of domination of the discussion drives away other voices and is bad for the collaborative process, clearly they should stop doing that." That's a lot of things to type, every time someone exhibits that behavior. When someone writes "per WP:BLUDGEON", they are saying basically what I said above, and their rationale is no less valid than someone who types out an entire paragraph. --Jayron32 15:19, 25 January 2023 (UTC)[reply]

Bludgeoning is more often than not encountered as a secondary "offence" together with with some other more serious abuse (often NOTHERE) and it is the combination that gives rises to sanctions. Bludgeoning would have to be laid on pretty thick to get sanctioned just for it. It's a bit like incivility, we shouldn't do it, but in the heat of the moment, we do, again rarely sanctionable, unless its ott. Both more frequent in CT areas. Idk, perhaps that is why it is an essay rather than a guideline.Selfstudier (talk) 23:13, 24 January 2023 (UTC)[reply]

Linking to essays when making a complaint about behavior is usually a bad thing. As a starting point, such essays might make a good vague point about behaviors to avoid, but lack the carefulness and specificity to be taken any further than that. So the "vague" is dangerous and easily abuse-able and when taken or used beyond that. They are used to paint someone as being bad without having to be specific abut the complaint. Lots of such linking is also used to give a false sense of credibility, wiki-expertise or authoritativeness to the complaint or disparagement. North8000 (talk) 20:10, 25 January 2023 (UTC)[reply]

  • I'd support the second option, though exact language would need to be drafted. BLUDGEON is a good essay, but it's written like an essay and not like a guideline. I think caution against bludgeoning would be a good guideline in two important senses: it reflects current admin practice—as admins are blocking users while citing the essay—and the community benefits from discussions where a multitude of voices are not crowded out by one loud one. Firefangledfeathers (talk / contribs) 08:05, 30 January 2023 (UTC)[reply]

The point of BLUDGEON, like many explanatory essays (those with the {{supplement}} tag), is to further elaborate on a specific problem/situation that is already covered by the existing policies and guidelines. That's why such essays don't have to be guidelines themselves: They contain nothing that is not already covered (in less detail) in the other guidelines and policies. A lot of essays like that exist (e.g. WP:BRD, WP:SNOW etc.) that have broad community support but that does not mean we should transform then into guidelines. That would be the kind of detailed WP:CREEP (another such supplement) that our policies and guidelines should strive to avoid. Regards SoWhy 09:42, 30 January 2023 (UTC)[reply]

Update policy editnotice

While not a change to a policy, it will have a change on all policy pages so I am posting here for visibility. The current editnotice comes off as a Wall of big text and no one really wants to read that. Editnotices are best when they have a title, bullet points of what to do and additional text for anyone who cares. I was inspired by {{Contentious topics/page restriction editnotice}} to update this editnotice and propose changing it to {{Policy or guideline editnotice/sandbox}}. I can't say I ever properly read the edit notice before since I never felt the need to read 4 lines for just an "attention", this change should help with that. Terasail[✉️] 15:10, 25 January 2023 (UTC)[reply]

I dig it. If no one is going to read it anyways, it's useful to have less material for them not to read. --Jayron32 15:13, 25 January 2023 (UTC)[reply]
They both seem fine, however the proposed one looks more like a "wall of text" that I'm not going to read then the current one to me. — xaosflux Talk 15:17, 25 January 2023 (UTC)[reply]
You think so? There is only one sentence which is the full width of the box (The final one), the other 4 are between 40%-60% of the width of the box, with 2 clear bullet points of what to do instead of having the instructions mixed in a text block. Terasail[✉️] 19:56, 25 January 2023 (UTC)[reply]
@Terasail, I've no opinion on the content, but I suggest dropping the icon. Those tend to display badly on narrow screens (e.g., mobile). Whatamidoing (WMF) (talk) 19:10, 26 January 2023 (UTC)[reply]
 Not done: No consensus for this being a positive change as proposed. Terasail[✉️] 01:53, 1 February 2023 (UTC)[reply]

Discussion at WT:CSD

There is currently a discussion about a possible new criterion at Wikipedia talk:Criteria for speedy deletion#New General Criteria?. At this point there are no concrete proposals, but several users have commented about problematic drafts that aren't obviously covered under any existing criteria. Any input would be appreciated. The Blade of the Northern Lights (話して下さい) 13:57, 26 January 2023 (UTC)[reply]

Adding an edit notice to standalone lists of organizations

I'd like to mass request an edit notice to lists of organizations. Specifically starting with all the pages listed in Category:Lists of high schools in the United States by state. Wikipedia:LISTORG indicates that for inclusion in a standalone list, an entry must have an existing page or be independently and reliably sourced. Thus, I have created Template:Editnotice listorg. I have also begun the process of removing entries from lists in the aforementioned category but am going to hold off for now until I discuss here and get approval. I'm hoping this will help cut down on link spam and cruft in these lists. Philipnelson99 (talk) 13:23, 27 January 2023 (UTC)[reply]

If the template needs changing, then that is perfectly fine by me. Philipnelson99 (talk) 18:46, 27 January 2023 (UTC)[reply]
It's unclear to me if you intend to put this template on the page or on the talk page? Jahaza (talk) 19:38, 27 January 2023 (UTC)[reply]
That's fine, I meant specifically the list pages included in Category:Lists of high schools in the United States by state, not the talk pages. Philipnelson99 (talk) 19:47, 27 January 2023 (UTC)[reply]
These lists don't have a notability criteria. So it might be better to add citations than to try to eliminate schools that fit the list inclusion criteria as cruft/link spam, because they're not really. There are databases that can serve as a source for a lot of schools. Jahaza (talk) 19:38, 27 January 2023 (UTC)[reply]
Does WP:LISTORG not apply here? If so then I agree that this isn't needed. But as it stands, I believe it does apply and therefore any schools added would need to meet the criteria specified in that policy. If I've misunderstood the policy that's fine. And if a blanket reference from an independent database suffices that's fine too, but it would obviously need to be referenced on the page. Most of these pages don't contain such references and if they contain references at all they are usually primary sources which don't meet the criteria in WP:LISTORG. Philipnelson99 (talk) 19:50, 27 January 2023 (UTC)[reply]
I think you're misreading WP:LISTORG? That guideline does apply, but it specifically says that members of a list don't have to be independently notable unless the list itself specifies that criteria. They just need to be cited. Since these lists don't have a notability criteria, any school added is not cruft as long as it can be verified. Meanwhile, it's better to add those citations rather than just blanked deleting the content. Jahaza (talk) 20:18, 27 January 2023 (UTC)[reply]
I am specifically looking at:

If the company or organization does not have an existing article in Wikipedia, a citation to an independent, reliable source should be provided to establish its membership in the list's group.

I acknowledge there is a sentence of policy before this that says the entry does not have to meet WP:NOTABILITY, but entries also to have meet the second part quoted above. Maybe removal of unsourced entries is too much but new entries should absolutely adhere to the policy. Philipnelson99 (talk) 20:26, 27 January 2023 (UTC)[reply]
Also take a look at this diff where I cleaned up a lot of entries that had no source at all. Adding sources to each one of these and verifying that they are indeed real schools would take an extraordinary amount of time. Not saying it's an unwarranted thing to do but lots of these articles are in this condition right now. Philipnelson99 (talk) 20:33, 27 January 2023 (UTC)[reply]
  • I would recommend you go ahead and undue your removal of non-linked public high schools from whatever states you have removed it from. You absolutely have failed to attempt to source these entries at all (where any sort of minor newspaper coverage at all would have sufficed). You removed Vincent Martinez from the list of California schools: here's a source right here [7]. You removed New Brockton High School: ditto [8]. It's one thing to remove external links, and it's another to make wide-sweeping removals which are in all honestly pretty terrible despite you acting in good faith.
Also, your new notice is wrong. You absolutely can add external links to lists per WP:ELLIST. For example, it would not be undue formatting it as:
School School District City Website
New London High School New London, Wisconsin New London School District Official website
Saying you cannot add links unless they are a reference is wrong. (If you don't revert, I will myself, as your WP:BOLD changes have been opposed.) Why? I Ask (talk) 10:32, 30 January 2023 (UTC)[reply]
Just to be clear, you've now reverted all the pages I changed back to their state from the other day. I'm not going to revert your changes because I'm not going to edit war but I still believet WP:LISTORG applies here and that the edit notice should be added which is the original proposal. Philipnelson99 (talk) 12:08, 1 February 2023 (UTC)[reply]
I believe you've misread WP:ELLIST.

External link sections are not prohibited at the end of stand-alone lists or at the end of articles that contain embedded lists. However, the lists themselves should not be composed of external links. These lists are primarily intended to provide direct information and internal navigation, not to be a directory of sites on the web.

— WP:ELLIST
There is a part in WP:ELLIST that says official links may be included if the list is of software or political candidates. I see no mention here of standalone lists of organizations. WP:LISTORG says that entries must be independently sourced or have existing Wikipedia pages. Philipnelson99 (talk) 12:46, 30 January 2023 (UTC)[reply]
No, that provision basically means that you can't have lists that read: As part as tables, yes they are allowed, using politicians and software as non-exhaustive examples. And the vast removal of easily verifiable information is just blatantly disruptive. Instead of taking the short time to look up the school on the internet, you removed them. These aren't facts that need careful studying, this is simply verifying that a school exists through local coverage. You have utterly missed the purpose of WP:LISTORG. And besides, you've also overlook the fact that many non-linked schools are mentioned at their school district as well. Why not create redirects? You're trying to create solutions to a problem that hardly exists or is easily fixable. Why? I Ask (talk) 13:35, 30 January 2023 (UTC)[reply]
I'm simply removing information that doesn't adhere to an existing policy, how is that disruptive? The information can always be added back in a way that adheres to policy in the future. Philipnelson99 (talk) 13:45, 30 January 2023 (UTC)[reply]
Dude, every state has a directory that lists every school. And guess what? Every single school you removed are on those directories. Each highschool is verifiably real and this meets the criteria for inclusion (private schools not included). It does no good for the encyclopedia to require citing that yes, "School A" is real when it takes a five seconds to search on the directories or Google. Why? I Ask (talk) 13:53, 30 January 2023 (UTC)[reply]
@Philipnelson99, you have misread ELLIST. @Why? I Ask is correct. It would be very strange if a guideline specifically recommended a table of "Candidate – Political party – Official website" but secretly meant to ban "School – Government agency – Official website".
(In case it matters, I wrote ELLIST.) WhatamIdoing (talk) 20:11, 31 January 2023 (UTC)[reply]
  • The key here is WP:NOTDIRECTORY. We shouldn’t list every high school in the country (any country)… only those that are notable on their own. Blueboar (talk) 14:32, 30 January 2023 (UTC)[reply]
    Exactly, I think the edit notice could be altered to include that information. Philipnelson99 (talk) 14:37, 30 January 2023 (UTC)[reply]
    That's an entirely separate issue than what this notice is trying to correct, and your own analysis of WP:DIRECTORY may also not be correct here. There's very few public high schools from each state without a page, and adding them to the list for completeness is not undue (per WP:CSC). (Most of those missing pages could be created relatively easily too). Or, as stated, they can be redirected to the school district.
    Basically, the edit notice is wrong, so I oppose it. And whether or not listing all public high schools of a state should be done is a completely separate discussion. Why? I Ask (talk) 15:11, 30 January 2023 (UTC)[reply]
    The edit notice is correct, it plainly states that no entries should be added that do not follow WP:LISTORG. Philipnelson99 (talk) 15:14, 30 January 2023 (UTC)[reply]
    No, I've already pointed out that the section about external links is wrong. Why? I Ask (talk) 15:15, 30 January 2023 (UTC)[reply]
    It's a policy, are you saying the policy is wrong? Philipnelson99 (talk) 15:16, 30 January 2023 (UTC)[reply]
    No, the policy is right. It says in other cases, such as for lists of political candidates and software, a list may be formatted as a table, and appropriate external links can be displayed compactly within the table. Yes, links can be included per the policy. Why? I Ask (talk) 15:18, 30 January 2023 (UTC)[reply]
    You still haven't said why exactly the edit notice is wrong. You shouldn't add entries without an independent source or unless they have a wikipedia page. Regardless of how easy it is to source an entry, it must be sourced or meet notability guidelines. Including an WP:ELOFFICIAL is an entirely different issue than providing an edit notice and removing non-adhering entries. Philipnelson99 (talk) 15:35, 30 January 2023 (UTC)[reply]
    Because the edit notice literally says Do NOT add external links. There's nothing prohibiting an editor from doing so. Otherwise, sure it's fine and correct. But the mass removal of easily citable information is not cool. Otherwise, every page that is unreferenced would be systemically deleted despite how easily notability can be proved. If you want to prevent editors from adding unsourced information in the future, great. Removing longstanding entries that you admit can be sourced but you've refused because "it takes too much time" (your words) is disruptive regardless of any policies. Why? I Ask (talk) 16:00, 30 January 2023 (UTC)[reply]
    @Blueboar, it's always been fine to make a list of non-notable subjects, or a mix of notable and non-notable subjects. (Weren't you around for the Pokémon wars 'way back in the misty dawn of the wiki?) I don't understand why you think a list of high schools must be limited to notable schools only. We wouldn't question a short list of high schools in the standard ==Education== section of a small city or a complete list of all schools in a school district (example), so why couldn't there be a longer list in a bigger area? "List of high schools in the United States" would be unwieldy, but "List of high schools in South Dakota" would be manageable. WhatamIdoing (talk) 20:16, 31 January 2023 (UTC)[reply]

Should non-admins be allowed to review unblock requests?

The following discussion is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.



There's an LTA that has a fondness for going through CAT:RFU and making bogus unblock reviews. [9] is an example. Discussion ongoing at Wikipedia:Edit filter/Requested#Illegitimate unblock request handling. As a policy issue, is there any reason to allow non-admins to review unblock requests? --jpgordon𝄢𝄆𝄐𝄇 20:01, 28 January 2023 (UTC)[reply]

I can't think of a legitimate case, outside of procedural closes after being under/TPA blocked after the review had been posted, or if it were added by another party. Realistically, this could still easily be done by admin.Lee Vilenski (talkcontribs) 20:22, 28 January 2023 (UTC)[reply]
Without the toolkit you cannot accept an unblock request and the only defensible declines would be procedural, but those are invariably of the type that could practically be done by a bot and so contribute minimally to the maintenance burden. Usually if someone who isn't a sysop starts to review unblock requests we ask them to stop even if the requests were certain to have been declined by anyone.
Be that as it may, if this is the LTA I think it is they will promptly find a new way to disrupt the unblock process or find a new internal process to disrupt, so don't get your hopes up too much even if an edit filter is implemented.
There's also the risk they will game the filters to perpetuate disruption by interfering with reverts by non-sysop RCP. I won't get into this for obvious reasons, but it's been done before and any filter would need to be constructed with due care. 74.73.224.126 (talk) 04:07, 29 January 2023 (UTC)[reply]
If a user makes a disruptive request, the request should be removed completely. Otherwise, the user should get a response from an account which is capable of unblocking - currently (and probably permanently) an admin account. Animal lover |666| 13:36, 29 January 2023 (UTC)[reply]
Wikipedia:Blocking policy#Unblocking says quite clearly that Any user may comment on an unblock request; however, only administrators may resolve the request (either declining or unblocking). Non-admins are not allowed to review unblock requests under the current policy as written. Chess (talk) (please use {{reply to|Chess}} on reply) 16:20, 30 January 2023 (UTC)[reply]
I can't see any reason why the policy should change to allow that. There are of course a few administrative tasks where a WP:NAC is welcomed when done appropriately. Unblock requests are certainly not one of them. The WordsmithTalk to me 02:05, 1 February 2023 (UTC)[reply]
  • As someone who has spend over a decade defending the rights of non-admins and demanding that we stop hassling them for doing work that doesn't need an admin to do; this is one of the rare cases where non-admins should not be handling the work. Non-admins can feel free to weigh in with their opinions, to comment, but the mechanical act of closing the unblock request should still be left to admins, because only admins can actually unblock them. From a policy point of view, I'm pretty sure this is already clear in policy, as noted above by Chess. Given that, I don't think we need to be careful in handling this troll in question. The edit filter is a good idea here. --Jayron32 16:05, 1 February 2023 (UTC)[reply]
The discussion above is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.

Create Wikipedia:Manual of Style/Years

Wikipedia:WikiProject Years currently has an issue where there are no agreed upon sitewide standard for such articles, and several users of the WikiProject are attempting to enforce ill-defined project standards on users inside and outside of the WikiProject. The most recent examples of this enforcement can be found at Talk:2022#RFC on the inclusion of Barbara Walters in Deaths (Result:) and Wikipedia talk:WikiProject Years#Defining international notability, but the WikiProject has been plagued with WP:OWNERSHIP issues for some time. I propose that the community as a whole create a guideline page at Wikipedia:Manual of Style/Years so an agreed upon standard can be implemented. Thebiguglyalien (talk) 18:45, 30 January 2023 (UTC)[reply]

There's not much issue regarding the style or layout of main year articles. The small minority of people wanting to remove all or the vast majority of the deaths are likely motivated primarily by wanting to avoid disagreements. The main issue is which events, births & deaths to include, which isn't a style issue. Jim Michael 2 (talk) 19:11, 30 January 2023 (UTC)[reply]
I would strongly disagree with this notion. More and more people don't like the current "absolute internationalism", and more and more people who attempt to add figures are categorized as "fans" and discriminately reverted/treated as such on previous talk page discussions. If Wikipedia is meant to be for our readers, it should include both what is norable AND what people want to see, albeit with some quality control to ensure that "the acquisition of Reichstag-Climbing Spider-Men, Inc. by the Angry Mastodon Corporation" isn't on there. InvadingInvader (userpage, talk) 20:46, 30 January 2023 (UTC)[reply]
I don't think an MOS page is the answer, as there isn't any issue with style or formatting, and MOS isn't the place for an inclusion criteria. there are no agreed upon sitewide standard for such articles is incorrect; as has been pointed out at Wikipedia talk:WikiProject Years#Defining international notability, there is indeed a sitewide standard for such articles, and it's WP:WEIGHT. I believe WikiProject Years has lost sight of the fact that these year articles are articles and thus are subject to the same WP:NPOV policy as all other articles. How do you determine what to include and what not to include? By looking at what WP:RSes include and exclude. There is no lack of RS "year in review" articles that can serve as sources. If an event is included in multiple RS "year in review" articles, then it's probably WP:DUE for inclusion in the Wikipedia year article. Same for deaths (if those continue to be included in year articles): there are plenty of RS "notable deaths of the year" articles. 2022 is supposed to be a summary of RSes that write about "2022 in review". Applying that logic to Barbara Walters (or any nominated inclusion) would easily resolve the question. The problem at WP:YEARS is people voting without reference to (and in some cases, such as Barbara Walters, obviously without even looking at) RSes. WP:YEARS should probably write an internal project guideline that reminds editors to look at RS and to apply WP:NPOV when deciding what to include and exclude. I don't think anything more is needed other than enforcing our existing content policies. Levivich (talk) 19:27, 30 January 2023 (UTC)[reply]
I generally agree with this approach regarding years articles, and it's what I initially tried a few months ago, but it was rejected by users that wish to maintain the year articles as they are. I made this proposal in the hopes that the broader community could provide input on what goes into these articles and how they're formatted, but really anything involving sitewide input would be great at this point, regardless of what conclusion it comes to. Thebiguglyalien (talk) 19:38, 30 January 2023 (UTC)[reply]
@Thebiguglyalien: Would you mind dropping a link to that discussion from a few months ago? I'd be interested in reading it. It sounds like you were on the right track then, and I agree that this merits broader community input and thank you for starting this discussion. Levivich (talk) 19:42, 30 January 2023 (UTC)[reply]
There was (and is) a lot going on. It got a bit heated and a bit spread out. The discussions in question are Wikipedia talk:WikiProject Years/Archive 15#Standalone lists – prose in year articles and Wikipedia talk:WikiProject Years#RfC on prose in year articles, but basically everything after that relates to the topic somehow. There's a novel's worth of arguments on the WikiProject talk page that basically goes in circles. I think the talk page there speaks for itself why I'm trying to reach out to the community. Thebiguglyalien (talk) 19:52, 30 January 2023 (UTC)[reply]
The large majority of year in review articles are country-centric. For example, those by the US media for 2022 would prominently include Depp v. Heard & the Will Smith–Chris Rock slapping incident as being among the most important events of the year, but not mention the 2022 Peshawar mosque attack. Jim Michael 2 (talk) 20:01, 30 January 2023 (UTC)[reply]
The large majority of year in review articles are country-centric is easily disproven. I typed in 2022 in review into Google, and here is what I found on the first two pages of results: AP, Reuters, CBS, The New Yorker, McKinsey, WSJ, NPR, CFR. Every single one of those is global in scope. CNN divides it into "US" and "international". There is no lack of RSes that do a year-in-review with a global scope. I didn't see the Peshawar mosque attack listed in any of those, although I did see other Pakistan news listed in many of those. It's almost as if ... wait for it ... the RSes don't think this mosque attack was one of the most important events that happened in 2022. We should follow those sources. 2022 Peshawar mosque attack does not appear WP:DUE for inclusion in 2022; it should be removed, if we followed WP:NPOV. Levivich (talk) 20:27, 30 January 2023 (UTC)[reply]
You have to be joking if you seriously think the Peshawar mosque terrorist attacks, which killed over 60 people and injured nearly 200, is less internationally notable than a slap on an awards show. How can you not see the issue with systemic bias? TheScrubby (talk) 20:39, 30 January 2023 (UTC)[reply]
How many of the links I just posted included the awards show slap? Therein lies my answer to whether it should be included in 2022. It's a simple algorithm: follow the sources. Levivich (talk) 21:00, 30 January 2023 (UTC)[reply]
Those many not, but many do. Many media sources portray the slap as one of the most important events of last year. Jim Michael 2 (talk) 22:00, 30 January 2023 (UTC)[reply]
If that were true, you'd have no problem posting some examples of "year in review" articles by reliable sources that list the slap. I've seen them in "2022 in entertainment" reviews--in which case, the slap is one of the biggest stories of 2022 in entertainment--but I have not seen them in the general overviews, such as the ones I posted above. Levivich (talk) 22:09, 30 January 2023 (UTC)[reply]
Global, maybe - but with a disproportionate amount of coverage from their own countries. You're helping to prove my point regarding many media sources' measure of importance. They print/broadcast what sells. We shouldn't follow their sensationalist, reader/viewer-chasing tactics. Jim Michael 2 (talk) 20:48, 30 January 2023 (UTC)[reply]
On Wikipedia, that sensationalist, reader/viewer-chasing media are called "WP:Reliable sources". I'd like to quote Snow Rise from a recent posting at WP:ITNC:

The only relevant evidence anyone should be advancing is that which is based in the volume and depth of coverage in sourcing, not idiosyncratic opinions directly from our editorial corps' mental processes as to why this subject is or is not "really important in the grander scheme of things". You think that "the media" is blowing the story out of proportion? Oh well, too bad--that is exactly as much an WP:OR argument here as it would be on the talk page for relevant articles. You know what we call "the media" in contexts such as this? WP:RELIABLE SOURCES--you know, the stuff we are supposed to be basing our content on, rather than our individual, personal opinions?

When editors overrule RS coverage with their own personal opinions about what is/isn't important, that violates our policies (like WP:DUE). Levivich (talk) 21:08, 30 January 2023 (UTC)[reply]
Most sensationalist outlets aren't reliable. Jim Michael 2 (talk) 22:00, 30 January 2023 (UTC)[reply]
AP, Reuters, CBS, NPR, WSJ, CNN are not reliable sources? Levivich (talk) 22:06, 30 January 2023 (UTC)[reply]
They are, but if they all omit the 2022 Peshawar bombing from their reviews of last year, it shows that they're selecting some of their stories based on what will gain the most readers/viewers rather than importance. Jim Michael 2 (talk) 22:19, 30 January 2023 (UTC)[reply]
Uh, "what will gain the most readers/viewers" == what is important to most people == what is most important. But aside from that tautology, if you're seriously suggesting that when reliable sources don't cover something you think should be covered, that means they're no longer reliable sources, then you should not be editing here at all. All of our core content policies are about not setting aside RSes and going with our own personal opinions instead. What you're suggseting is way, way outside what is allowed on this website; it's quite literally the very opposite of WP:V, WP:NOR and WP:NPOV. Levivich (talk) 22:25, 30 January 2023 (UTC)[reply]
International notability is, and ought to continue to be, the standard for inclusion on the main Year articles, which are meant to be a summary of the most important events around the world rather than just one or a handful of countries. I reject very strongly the dismissal of the (very fair) arguments against the inclusion of Walters by Levivich, and I reject any move that would in any way help perpetuate systemic Americentric bias. What ought to be done is to look at our current international notability standards and make necessary improvements (there is always room for improvement) to address the concerns of some users, as well as to work on a criteria for entertainment figures (like the criteria already in place for political and sports figures). We don’t need to treat dandruff with decapitation, and while it’s not a perfect system I think basing inclusion on substantial international notability for the main international year pages is the best way to go. Not to return to how things used to be, which clearly was not working and was leading to Americentric bias as described in the WikiProject talk page, and I oppose any changes that would in any way help perpetuate this systemic bias, including the notion of relying solely on international media sources (if that was allegedly how most inclusions on the main yearly pages were decided, then the fact that it led to overwhelming Americentric bias where all things American were added with far less scrutiny than its international counterparts, and that “RSes” strongly favour America/the Anglosphere over the rest of the world, then clearly it wasn’t working and that change was absolutely necessary). TheScrubby (talk) 20:27, 30 January 2023 (UTC)[reply]
My belief is that international notability should be preferred, albeit there should be more leeway for the inclusion of notable domestic events with substantial international coverage. Events that currently aren't included or are "borderline included" (as defined by previous discussions on Main Year Article talk pages) would be January 6, the recent events in Brasilia, events which were domestic but contributed to inciting a major global movement (such as the Killing of Oscar Grant being a key inciting incident to the BLM movement), well-covered international protests like Mahsa Amini and the US-Canadian Freedom Convoy, and at a stretch, Dobbs v. Jackson.
In terms of the implementation of an MOS guideline, I'd like to only see it implemented as a last resort. I think that Thebiguglyalien makes quite a few good points on WikiProject Years exhibiting an unacceptable amount of ownership behavior, even if slightly over the line. I'm not exactly keen on the abolition of the WP entirely, but it's something that should be looked at. My view on the problem is that it seems to be a lack of compromise between the more exclusionist editors and the more inclusionist ones. No one seems to be willing to work together on some key inclusions, and while we're all mostly behind excluding stuff like Gabby Petito on Main Year articles, too many disputes have caused controversy. The Deaths section in particular generates a lot of heated debate and wasted time that could have been spent on and off wiki, leading me to support the complete or near-complete removal of deaths entirely and provide a hatnote to the Deaths article (though I have suggested as a compromise that these discussions judge the inclusion of sidebar photos in deaths). InvadingInvader (userpage, talk) 20:58, 30 January 2023 (UTC)[reply]
People add insufficiently notable things to main year articles every day. We need more regular editors to remove them.
The killing of Grant was domestic, 99% of people outside the US haven't heard of it & it shouldn't be included. If it were important to BLM's founding, it wouldn't have taken them 4 years. We don't include things on the basis of them being one of the things that lead up to a group being founded; there are many such events.
You're the only editor who wants Dobbs included. Those protests had a death toll of zero, were overwhelmingly domestic & had no effect. Protests happen all the time; we only include the most important ones. Jim Michael 2 (talk) 21:25, 30 January 2023 (UTC)[reply]
People add insufficiently notable things to main year articles every day. Is there a notability guideline for Year articles? My guess from WP:NOTEWORTHY would be "no", but curious if there's something I've missed. we only include the most important ones. Who decides that? Our sources, or you? —Locke Coletc 22:54, 30 January 2023 (UTC)[reply]
Yes, decided over many discussions regarding that. Trivia that's frequently added to main year articles includes local sports events & clubs, openings of shopping malls, local weather records, concerts, releases of singles & albums, premieres of plays & films, the start & end dates of radio & TV shows as well as broadcast dates of 'particularly important episodes' of them, when celebrities met their partners, married & divorced them etc. We need more regular editors removing these things. Jim Michael 2 (talk) 23:10, 30 January 2023 (UTC)[reply]
Yes, decided over many discussions regarding that. If those "discussions" weren't involving our reliable sources, then it sounds like a group of editors were off engaging in WP:OR. The way to keep local sports events & clubs, openings of shopping malls, local weather records, concerts, releases of singles & albums, premieres of plays & films, the start & end dates of radio & TV shows as well as broadcast dates of 'particularly important episodes' of them, when celebrities met their partners, married & divorced them etc. out of year articles is to look to whether or not reliable sources provide coverage of the items being added in their year-end roundup articles. —Locke Coletc 23:19, 30 January 2023 (UTC)[reply]
All entries are required to be backed by reliable sources. You don't understand the problem: every day fans add what they want to add to main year articles regardless of the rules/guidelines/inclusion criteria. There are far too few regulars to remove them. Loads of people love to add the person or event that they want to promote, such as a 'really important' episode of Friends, the highest temperature recorded in Brisbane in the 20th century, a band's 'crucial' second album, a celebrity wedding, the biggest shopping mall in Calgary reopening after renovations, the opening of a play on Broadway & the births & deaths of various people who have no international notability. Such things are often on main year articles for months or years. They clearly don't belong there. If people put half the effort into removing them rather than starting discussions like these & RfCs for people who've already been excluded by consensus due to their insufficient international notability, main year articles would be a lot better. Jim Michael 2 (talk) 00:47, 31 January 2023 (UTC)[reply]
All entries are required to be backed by reliable sources and starting discussions like these & RfCs for people who've already been excluded by consensus are problematic when being used to discuss the same problem. We don't "exclude content by consensus". We follow what our reliable sources on the matter state, using Wikipedia's voice to distill that information down to what is encyclopedic. I'm not saying there shouldn't be some kind of standard establishing how many reliable sources must cover the topic being debated to merit inclusion, but the way it's been described here rings of people just engaging in a long-form version of WP:ILIKEIT (or the contrary) which is original research.. but the regulars in these articles wishing to exclude Barbara Walters from a list of deaths for 2022 tells me nobody was checking the sources when that decision was made. —Locke Coletc 01:28, 31 January 2023 (UTC)[reply]
The significance outside of the United States of Walters’ has been greatly exaggerated by many users arguing in favour of inclusion. Whatever the final outcome of the discussion may be, I think it’s important that the discussion took place to begin with. It’s important that we scrutinise potential inclusions and that those with questionable levels of international notability who are much more significant domestically within their own country are discussed rather than automatically included (complete with image, as some users attempted to include initially - and we never include images of figures who were borderline inclusions or were only included after heavy debate necessitating an RFC) without question because of the country they happen to come from. After reviewing the discussion and what was brought to it, it’s clear that Walters did have some international influence within her field of journalism, although I am not convinced that it’s sufficiently internationally notable enough (certainly not for automatic inclusion) and that she is still predominately a domestic figure mainly relevant to the United States. I can absolutely guarantee that had Walters been from any other country, and had achieved everything she did within any other country other than the United States, the discussion over her inclusion would have been very different - and almost certainly considerably shorter. TheScrubby (talk) 01:45, 31 January 2023 (UTC)[reply]
@TheScrubby: Which reliable sources do you consult when arriving at a conclusion about including/excluding an entry, e.g. Walters? Levivich (talk) 02:04, 31 January 2023 (UTC)[reply]
The person's international career, international awards etc. Jim Michael 2 (talk) 02:40, 31 January 2023 (UTC)[reply]
Inclusion/exclusion is based on the person's accomplishments? Levivich (talk) 03:29, 31 January 2023 (UTC)[reply]
For main year articles, yes. For example, Olympic medals for sportspeople, international acting awards for actors, playing international concerts &/or having international chart hits for musicians, being a head of state/gov for politicians. Those are substantial international notability. Jim Michael 2 (talk) 03:47, 31 January 2023 (UTC)[reply]
And this has been the consensus at WP:YEARS since mid-2021? Levivich (talk) 03:58, 31 January 2023 (UTC)[reply]
We have a criteria in place for political figures since mid-2021 (which you can read on the FAQs sections of each of the recent years pages), after a long discussion on the matter and with a consensus that has been backed up by successive actions by virtually all regular users since. What were things like prior to this consensus? Minor, purely domestic US politicians such as John B. Anderson, John Dingell and Ralph Hall were included for years without discussion or scrutiny, whereas their international equivalents (or indeed politicians from outside the US in general who actually held senior ministerial positions) were summarily questioned and removed. Having one set of standards for American figures, and another for everybody else. As for the rest, multiple discussions over the successive main year Talk pages have concluded in favour of including individual (not team) Olympic gold medalists. Entertainment figures are still in need of a criteria written up as a guideline for inclusion, although there is broad agreement on including core, essential members of internationally notable bands; musicians who had more than one major international chart hit; and recipients of major international film awards for directors and actors. TheScrubby (talk) 04:40, 31 January 2023 (UTC)[reply]
Some of the criteria have had consensus for longer than that, including automatic inclusion of politicians if they are current or former heads of state/gov, and exclusion of all others with the exception of those who nevertheless have substantial international notability, such Donald Rumsfeld. Jim Michael 2 (talk) 11:18, 31 January 2023 (UTC)[reply]
The bar here is international notability, and consensus on the main year pages is that international media sources do not, and should not, automatically equate to international notability. International notability has been the basic standard since mid-2021, due to severe issues with systemic bias before that where figures and events from one country were included with little to no scrutiny while figures/events from everywhere else were treated with a different standard entirely. If you’re looking for a concise definition of international notability as has been applied by multiple regulars over the last couple of years on the main international yearly pages, one was written a few days ago on the WikiProject Years talk page. TheScrubby (talk) 00:00, 31 January 2023 (UTC)[reply]
...has been the basic standard since mid-2021...: This "standard" has no Wikipedia Project advice page, let alone evidence of wide community support e.g. RfC. —Bagumba (talk) 01:08, 31 January 2023 (UTC)[reply]
I'm just gonna direct your attention to WP:LOCALCON: Consensus among a limited group of editors, at one place and time, cannot override community consensus on a wider scale. For instance, unless they can convince the broader community that such action is right, participants in a WikiProject cannot decide that some generally accepted policy or guideline does not apply to articles within its scope. WikiProject advice pages, how-to and information pages, and template documentation pages have not formally been approved by the community through the policy and guideline proposal process, thus have no more status than an essay. We will not be supplanting WP:V, WP:NOR or WP:NPOV with whatever flavor of the month a WikiProject has chosen. A consensus there has no more status than an essay. —Locke Coletc 01:57, 31 January 2023 (UTC)[reply]
Why are you trying to defend a system that clearly perpetuates systemic bias, and attempting to delegitimatise a system for main international yearly pages that bases inclusion on international notability rather than what is notable in the United States, and has been in place and supported by regular users since mid-2021 (which, by my estimation, is a little more than a “flavour of the month”)? Why even bother having the main international yearly pages or the Year In Country pages if 2023 becomes filled with domestic American events/figures and to hell with their international counterparts? The way things were before mid-2021 was utterly broken; figures and events were arbitrarily added and so many minor figures with scant international notability were included without any question or debate, oftentimes while equivalent figures were excluded. You lot continue to cite “reliable sources” as the be-all and end-all, and yet you seemingly choose to blindly ignore that even that system isn’t perfect and that such a system for these pages contribute to systemic bias in favour of figures from the Anglosphere, especially the United States. None of you have even attempted to come up with an alternative to either the current system or the broken Americentric system from before, and are instead just blindly going on about “reliable sources” without any acknowledgement of the flaws of relying solely on that system is for inclusion on the main international yearly pages. Frankly, it’s not helpful or constructive at all, and if relying solely on “reliable sources” was how inclusions were decided prior to mid-2021, either that was appallingly applied in practice, or that it clearly wasn’t working at all and that it led to lists that was overlong (especially with domestic figures/events) and easily exceeded the recommended maximum size for a Wikipedia article, then all I can say is that the definition of insanity is doing the same thing that’s clearly not working over and over again. TheScrubby (talk) 02:15, 31 January 2023 (UTC)[reply]
Why are you trying to supplant what our reliable sources indicate are the notable events for a year with whatever you and your peers choose using whatever "standard" happens to be at play in the moment? If you want to change the world, go join a newsroom or create a podcast or whatever else tickles your fancy to undo the injustices of systemic bias, but Wikipedia is not a battleground. The only world-changing that ought to be going on here is making knowledge free for all. Wikipedia documents the world as it is, good and bad. —Locke Coletc 02:26, 31 January 2023 (UTC)[reply]
You mean the standard of removing systemic bias by applying the same standards for events/figures of every country, and including events/figures of substantial international notability for the main international yearly pages? You’re clearly ignoring all of the issues which led to the adoption of the international notability criteria completely. How is it at all controversial to include people/events based on notability that goes beyond one country/region; being based on major international awards; representing their country on the international stage for major international competitions/conferences/organisations; and having an impact of international consequence based on the work in their relevant field beyond just their home country/region? Which, believe it or not, is a system that is far less arbitrary than how things were beforehand - and hey, this was all already de facto applied to figures/events of most non-Anglosphere countries, but certainly not for American figures where there has always been one standard for them and another for everyone else. Sources are useful for the main international year pages as a secondary factor for inclusion, and in helping back up all of the above, but not for arbitrarily using it as the one and only, be-all and end-all standard for inclusion, with all the flaws that comes with it. TheScrubby (talk) 02:36, 31 January 2023 (UTC)[reply]
Furthermore, before you try and deny/put your fingers in your ears about systemic bias here, how on earth can you possibly justify the fact that the main international yearly pages use the American mdy date system when the overwhelming majority of countries internationally use dmy? Even InvadingInvader, who started the RFC on Walters (and is in favour of her inclusion) and who thinks we should be more flexible on the international notability criteria, acknowledges and accepts that there’s issues with Americentric bias, and that the previous Americentric system for the main yearly pages was deeply flawed and untenable. Refusing to address these issues and going on instead about “reliable sources” is little more than a reactionary cop-out. TheScrubby (talk) 02:45, 31 January 2023 (UTC)[reply]
Do you understand what main year articles are for? Do you understand why the large majority of things that are on year by country articles aren't on main year articles? Jim Michael 2 (talk) 02:40, 31 January 2023 (UTC)[reply]
Because we violated WP:SUMMARY? Thebiguglyalien (talk) 02:47, 31 January 2023 (UTC)[reply]
We didn't violate anything. Main year articles are for international events & people. Year by country articles are for each country's. For example, Depp v. Heard is domestic, so it's on 2022 in the United States but not 2022. Main year articles include links to the subarticles. Jim Michael 2 (talk) 02:57, 31 January 2023 (UTC)[reply]
We don't need an RfC for everything. International media coverage doesn't grant international notability. If it did, James Michael Tyler would be regarded as having very high international notability & therefore be included. Jim Michael 2 (talk) 02:22, 31 January 2023 (UTC)[reply]
Jim, could I ask you to stop it with the "We don't do this" arguments? We're discussing what we should do, and when what we do is on the table, it doesn't seem productive to the discussion to say "what we don't do" and "what we do do" without providing a reason to change aside from appeal to tradition...in this discussion especially, this is like "We don't do things because we said so" (possible WP:OWN). InvadingInvader (userpage, talk) 21:41, 31 January 2023 (UTC)[reply]

section break 1

Copying and pasting from the WikiProject years as well as the FAQs section, for easy access on this discussion here’s international notability as how we have defined and implemented it on the main international Year pages since mid-2021.

International notability is:

  • Notability that goes beyond one country or one region/a handful of countries, and is not merely domestic.
  • Being the recipient of multiple major international awards, rather than just awards that are domestic and mainly given to those from their native country. In other words, representing their country on the world stage
  • Having an impact with their work in their relevant field beyond their home country/region, and are of international consequence; being an essential, central member of an internationally notable group; or achieving international notability beyond their most famous work (in other words, for example with actors they would be internationally recognised in their own right rather than say, being known as “x character in y franchise” and only being known internationally by fans)

International notability is not:

  • Measured by international media sources/obituaries. They’re essential on this Wiki for who is notable enough to warrant an article, or for other lists, but not for a page that is focused on the most internationally notable figures and events. Furthermore, using media sources runs the risk of perpetuating systemic bias, particularly in favour of figures/events from the Anglosphere over figures/events from the non-English speaking world.
  • Measured by the number of Wiki language articles or article views. Anybody can create language articles, including hardcore fans of celebrities, as is prominently the case with Corbin Bleu. Page figures can be an indicator of what is trending at the time, but cannot be used as a factor for determining international notability - which is not a popularity contest; nor are the international yearly articles meant to resemble tabloid papers (so we wouldn’t include trendy human interest stories that is not of lasting international significance).
  • Measured by the number of fans somebody may have internationally.
  • Having one set of standards for figures/events from one or a handful of countries, and another set of standards for figures/events from everywhere else. What’s notable in one country is not automatically notable elsewhere, and we would not include figures/events if their international counterparts are excluded. Americentrism or any other form of systemic bias is firmly repudiated.
  • Including people on the basis of quotas, tokenism or positive discrimination.

As a bonus, here is the political criteria we have in place, in line with the international notability requirement for the main international yearly pages:

1. Automatic inclusion: Heads of government/state (prime ministers, presidents, chancellors, governor-generals, etc.), figures regarded as central founding fathers of their nations, and heads of major intergovernmental organisations (such as the Secretary-General of the United Nations, President of the European Commission, etc.)
2. Considered for inclusion: Politicians who served as Foreign Affairs Minister, Secretary of State, Foreign Secretary, etc. whose actions were internationally notable and/or consequential.
3. Case-by-case basis: any other politicians (including deputy heads of government/state and opposition leaders/presidential candidates who failed to win an election) if they are internationally notable and/or consequential for reasons other than just holding the title of their office (such as those who won major international prizes such as the Nobel Prize)

And the sports criteria:

Tier 1: Association football, Cricket, Golf, and Tennis - as well as individual (not team) gold medalists of international sports competitions such as the Olympic Games. Include figures if at least globally known within the sport's culture.
Tier 2: Baseball, Ice hockey, Rugby football, Boxing, and Basketball. Include figures if globally known beyond the sport, and include the most notable of the sport's world ambassadors or most widely-successful promoters.
Tier 3: National footballs (Australian rules football, Gaelic football, Gridiron football), Lacrosse, Field hockey, Professional wrestling, and everything else. Generally exclude with exceptions made only for the most outstanding, arguably “once in a generation” type figures, equivalent to Pelé and Lionel Messi for soccer, or Don Bradman and Sachin Tendulkar for cricket - though in the event of inclusion they will not be prioritised for an image.


Now, tell me how these criteria for inclusion on the main international yearly pages are inadequate to the point where we need to repudiate it entirely and preference a system that for years allowed for systemic Americentric bias for pages that are meant to be the main international year pages. Certainly there’s always room for improvement, and I hope that users can get cracking on an entertainment criteria especially. But I’d honestly like to know why some users here are so dogmatically against any notion of using an internationally notable criteria where we apply the same standards to figures/events from every country without exception, and would rather a system that allows for systemic Americentric bias. TheScrubby (talk) 06:12, 31 January 2023 (UTC)[reply]

Articles about years, like 2022, are still WP:ARTICLEs, and subject to the same policies whether written in WP:PROSE (preferred), as a list, or a mix. WP:Stand-alone lists says: Being articles, stand-alone lists are subject to Wikipedia's content policies, such as verifiability, no original research, neutral point of view, and what Wikipedia is not, as well as the notability guidelines.
When determining whether certain content should or should not be included in a year article, treat each aspect with a weight proportional to its treatment in the body of reliable, published material on the subject (WP:PROPORTION). Keep in mind that, in determining proper weight, we consider a viewpoint's prevalence in reliable sources, not its prevalence among Wikipedia editors or the general public. (WP:WEIGHT). When writing about a topic, basing content on the best respected and most authoritative reliable sources helps to prevent bias, undue weight, and other NPOV disagreements (WP:BESTSOURCES). And as has been pointed out, unless they can convince the broader community that such action is right, a Wikiproject cannot decide that some generally accepted policy or guideline does not apply to articles within its scope (WP:CONLEVEL).
Articles about years are summaries of reliable sources about those years, and what events are included should be determined by the event's treatment in the body of reliable sources about the topic (about the year for main year articles, or about the year-in-topic for sub-articles). To vary that would require a widely-advertised WP:RFC, which I don't think has happened yet (and I don't think would be successful). Levivich (talk) 06:19, 31 January 2023 (UTC)[reply]
Levivich, Locke Cole: I've tried to explain this to them nearly a dozen different ways. These two (and a few other users on the WikiProject talk page) have been giving the same sort of responses for months. Thebiguglyalien (talk) 06:39, 31 January 2023 (UTC)[reply]
And yet nobody has given a sufficient response on how to deal with systemic Americentric bias as we have explained again and again, and which especially plagued the main yearly pages before the adoption of the international notability criteria. I’m not even talking American figures making up more than those from other countries; I’m talking cases where minor domestic American figures (especially B-list and C-list actors, and politicians who weren’t even senior figures) are not scrutinised the same way as their international counterparts, and with nobody demanding “reliable sources” to demonstrate their suitability for inclusion. I’ve given example after example on the WikiProject Years Talk page, and all you guys have is “reliable sources this, reliable sources that”.
A potential compromise would be to retain the international notability criteria, but to rely on “reliable sources” from outside their native country and from more than one language source, in order to demonstrate the subject’s international notability and importance/influence beyond their native country/region - with the burden of proof on those arguing for inclusion. So long as we don’t go back to how things were before mid-2021. TheScrubby (talk) 06:52, 31 January 2023 (UTC)[reply]
"When writing about a topic, basing content on the best respected and most authoritative reliable sources helps to prevent bias". The WP:ONUS is always on those arguing for inclusion. Arguments for and against inclusion of an entry in a year article should be based on the entry's prominence in reliable sources about the year. One of the things WP:YEARS could do is to compile a list of the best respected and most authoritative reliable sources about years, from around the world. For the current year, it's the world's best news media. For recent years, annual retrospectives. For older years, scholarship like history books and articles. Levivich (talk) 07:19, 31 January 2023 (UTC)[reply]
And yet nobody has given a sufficient response on how to deal with systemic Americentric bias as we have explained again and again Because we're here to write an encyclopedia, not to try to right great wrongs: You might think that Wikipedia is a great place to set the record straight and right great wrongs, but that is absolutely not the case. While we can record the righting of great wrongs, we can't actually "ride the crest of the wave" ourselves. We are, by design, supposed to be "behind the curve". This is because we only report what is verifiable using secondary reliable sources, giving appropriate weight to the balance of informed opinion. That being said, there is a countering systemic bias WikiProject, and from a brief look at that, it seems far more reasonable than the ideas pushed here (ignoring reliable sources and supplanting that with the will of a handful of editors who cherry pick things to bring their own sense of "balance" to this very niche-specific subject). Countering systemic bias in how we produce articles? Sure. Making sure we try to be inclusive of international/gender/racial views/subjects? Absolutely. But we don't get to start second-guessing reliable sources to make that happen. —Locke Coletc 08:10, 31 January 2023 (UTC)[reply]
  • I have been a frequent editor for over 13 years and an active administrator for over five years. Until today, I had never once heard of the concept of "international notability". In my view, there is no such thing and we do not have flavors or subcategories of notability. Of course, we debate borderline cases but in the end, we need to decide that topic A is notable but topic B is not notable. That's it. I am all in favor of efforts to deal with systemic bias, but repeated arguments that Amanpour is more "internationally notable" than Walters, as if that was actually a thing, is laughable. Both are notable, but Walters died and all sane people hope that Amanpour lives for many years to come. Inclusion criteria need to be rooted in what reliable sources say about the topic, not on the whims of a handful of editors at a Wikiproject. The notion that sources that enjoy project-wide consensus as reliable sources are somehow unacceptably sensationalistic in their end-of-the-year coverage, and should be rejected by a small clique of editors at this group of articles is utterly bizarre. We summarize what reliable sources say about topics, nothing more and nothing less. Period, end of story. We have no "super-editors" who are self-deputized to call coverage in reliable sources unacceptably sensationalistic. Gain projectwide consensus to deprecate such sources, if you want to try that, but accept them as reliable until consensus changes. I do not feel strongly one way or another about the inclusion of Barbara Walters in particular but analogizing her death to local temperature records or the opening of a shopping mall is deeply inappropriate. I am an American but have no wish to exaggerate my country's importance. At the same time, I cannot ignore efforts to diminish the importance of the United States on the world stage, and especially its role in the Anglosphere. There is a serious problem here, and Levivich has written some very perceptive things above. Cullen328 (talk) 07:47, 31 January 2023 (UTC)[reply]
There clearly are a significant minority of people & events which are significantly international & are therefore eligible for main year articles. Domestic events are for the many year by country articles. We don't wait for end of year coverage; these articles are edited every day. I'm not saying that all the insufficiently notable events are equal or similar; I listed many of the types of things that are wrongly added to main year articles. Being anti-Americentric isn't being anti-American. The US & to a lesser extent the rest of the developed Western world is greatly overrepresented on main year articles & WP in general. Jim Michael 2 (talk) 11:18, 31 January 2023 (UTC)[reply]
It really is sad. The mentality operating here is exactly the sort of activist WP:BATTLEGROUND stuff that we try so hard to avoid. If the consensus that previously existed on these pages is being challenged, and that consensus has not been challenged for multiple years, that does not mean that the people challenging it are wrong nor does it mean the consensus is here to stay forever more. WP:CCC is specifically part of established policy for a reason. At Wikipedia:In the news, we ourselves have been dealing with a similar phenomenon, when it comes to international notability being pushed as a "criterion" and thus used to challenge stories that might be primarily based out of the U.S. or the U.K. As is noted in WP:ITNCRIT, it's not required for a story to gain consensus. And as Cullen328 has so eloquently pointed out, nowhere on Wikipedia does international notability exist as a policy-based criterion for excluding or including anything from anywhere. There might have been a local consensus at one point in time, and there might still be concerted efforts through WP:SYSTEMICBIAS to add more events and persons to Wikipedia that might have been excluded in the past due to race, gender, ethnicity, etc. But that is not a reason to remove events and persons that happen to be from the United States if they demonstrate notability, which in these aforementioned cases, they clearly do. 🌈WaltCip-(talk) 13:25, 31 January 2023 (UTC)[reply]
Sad is that the criteria since mid-2021 is more biased than the media. We won't reduce systemic bias by ranking golf and tennis, two very white sports, as Tier 1, while ranking basketball--a more popular sport played overwhelmingly by people of color--as Tier 2, and ranking professional wrestling--which isn't a sport at all--as Tier 3 along with "national footballs"--which gridiron football is not, because it's played internationally. This is the sort of sword-skeleton theory we get when we trade reliable sources for the opinions of editors. Instead of reducing bias, we make it worse. Trading Americentrism for Eurocentrism won't help us learn about Asia, Africa, or the rest of the world. Levivich (talk) 14:59, 31 January 2023 (UTC)[reply]
Golf and tennis are two of the most internationally played sports, and so many of the world’s most famous players from both have not been “white”. Basketball is in the right tier - it’s a sport that has grown internationally in recent decades but is still not on the level of the Tier One sports. Gridiron football is a national football of just one country - like Australian Rules or Gaelic football, it’s played outside their main country but not widely played. Tier 3 is absolutely the most appropriate category for them. Professional wrestling was only included (and included in the third tier) because we’ve had issues over the years of super fans attempting to add professional wrestling figures with scant international notability. None of this was on the basis of bias of any sort, and was decided on the basis of how internationally notable and played each sport is. Any assertion that there’s Eurocentrism of any kind on the main yearly lists are complete nonsense - we treat figures from Europe no different from Africa, Asia, the Americas, or elsewhere. TheScrubby (talk) 22:47, 31 January 2023 (UTC)[reply]
I'm no expert on the subject, but I don't think promoting golf and tennis is the best way to fight systemic bias. Levivich (talk) 22:56, 31 January 2023 (UTC)[reply]
Both sports are among the most widely played internationally and in terms of the number of countries that they enjoy popularity in. It would be systemic bias to put gridiron football any higher when the sport is barely played outside the United States. TheScrubby (talk) 23:01, 31 January 2023 (UTC)[reply]
Systemic bias is about more than just counting the number of countries. Pro golf and tennis are among the least diverse sports in the world, and rather famously so. Levivich (talk) 23:12, 31 January 2023 (UTC)[reply]
International notability is notability that goes beyond one country/one region. It’s that simple. Figures and events with domestic notability go in Year In Country pages, while figures and events with substantial international notability go in the main international year pages. TheScrubby (talk) 23:06, 31 January 2023 (UTC)[reply]
  • With the caveat that I am not a member of WikiProject Years nor particularly active on years articles, I very much support efforts to fight systemic bias in this area by ensuring that articles have an appropriately international focus. We are an encyclopedia for all of humanity, not just the Anglosphere. It is entirely foreseeable that year articles, given their broad scope, will have a tendency to accrue Anglocentric cruft, and it is absolutely our role to push back against that. {{u|Sdkb}}talk 16:32, 31 January 2023 (UTC)[reply]
    The thing is, nobody disagrees with this, which makes this a bit of a straw man argument. We all agree about systemic bias, WP:SYSTEMICBIAS was written in 2006. The question isn't whether systemic bias exists or doesn't, nor is it whether we should fight it or not, the question is how. Levivich (talk) 16:35, 31 January 2023 (UTC)[reply]
  • It seems like the underlying issue here is that additions of Ameri/Anglo-centric information to year articles outpace the addition of other material, due to English Wikipedia's own demographic. The solution IMO would be to proactively add more information about events outside that sphere, or to slap a {{globalize}} template on the article if it's still a work in progress. In principle, I think it would be theoretically possible to come up with guidelines for inclusion that are more specific than "does the topic meet article notability guidelines" or systemic curbs on overrepresenting a country (e.g. to add a WP:LASTING-style consideration, or adding it for overrepresented regions), but in practice the range of material that is viable to include in a year article is so broad that establishing such standards would likely be quite difficult. signed, Rosguill talk 16:42, 31 January 2023 (UTC)[reply]
  • International notability is not...Including people on the basis of quotas, tokenism or positive discrimination: But that is exactly what is being done by merely seeing too many Americans and creating rules to arbitrarily boost subjects with mentions in lower-population countries while ignoring WP:WEIGHT. Any "solutions" of systemic bias need to have a litmus test of addressing the lack of women, and subjects from high-population countries like China and India. This otherwise has a Eurocentric male tilt.—Bagumba (talk) 18:45, 31 January 2023 (UTC)[reply]
    The issue hasn’t specifically been “too many Americans”. Even with the international notability criteria in place, Americans still have more representation than figures from any other country. We acknowledge that due to its population size and the influence of American culture internationally, they would have a higher percentage of people with sufficient international notability. No, the issue has always been the treatment of American figures with insufficient international notability - the likes of John B. Anderson and James Michael Tyler, people on those levels of notability. Figures who were not internationally significant or even major figures in their respective fields, whose international counterparts would never have been included by most users and, if they were included, they would be immediately scrutinised and removed - something that prior to 2021 simply didn’t happen in a major way with American figures on that level. And year death entires prior to 2021 were flooded with precisely these kind of entries. The inclusion for years of an image of the domestic politician and failed Presidential nominee (which with most countries their Opposition leaders and failed Presidential candidates are rarely ever included) John McCain in August 2018 over the Indian Prime Minister Atal Bihari Vajpayee (who was internationally consequential with the actions he took in office as head of government of the country with the world’s second-largest population - and is the world’s largest democracy) is absolutely inexcusable and a damning indictment on the Americentric bias of this Wiki. Don’t have one standard for figures/events from one country, and another for figures/events from everywhere else. TheScrubby (talk) 22:57, 31 January 2023 (UTC)[reply]
Reading through all the linked to discussions, it really does seem like a major ownership issue from the people of said Wikiproject that are attempting to internally make a rigid claim about inclusion notability without involving the wider community (who quite possibly would disagree). And I do think a good point is given above that we should be fixing systemic bias by adding more things to said articles to cover underrepresented regions and groups. But attempting to restrict or remove rather blatantly notable subjects in order to facilitate that is a nonstarter. As a comparison, the method for members of WiR like myself to increase the coverage of women on the project is to make more articles on women, not to delete articles on men. SilverserenC 01:16, 1 February 2023 (UTC)[reply]
How can it be WP:OWN when consensus has existed since mid-2021 among all regular contributors on the yearly pages that international notability is the bar for inclusion (with disputes mainly being over the extent of which the standard should be applied), and that domestic figures/events go into Year In Country articles? Consensus has always been respected on these pages - there are many instances time and again where certain regular users for one reason or another has always conceded to consensus even if they disagree with it - to use an example involving myself, I conceded to consensus when Norm Macdonald was included even if I argued against inclusion on the basis that he lacked substantial international notability as an actor, and that his notability was mainly restricted to North America (and I’m sure @Jim Michael 2: or any of the other regulars can point to many examples where they conceded to consensus for the inclusion of certain figures even if they personally disagreed). As for “attempting to restrict or remove rather blatantly notable subjects in order to facilitate that”, please read my comment directly above yours. TheScrubby (talk) 01:39, 1 February 2023 (UTC)[reply]
Your response presupposes four things:
  • 1. That the consensus which existed since mid-2021 regarding international notability still has standing, meaning WP:CCC does not apply,
  • 2. That "certain regular users" should have the final say on what gets posted into the Year articles, except by the goodness of their hearts in which they concede, which is the definition of WP:OWN,
  • 3. The failure by users to nominate international counterparts automatically means that we should continue excluding notable figures forevermore on the basis of avoiding Americentrism, which is not what WP:CSB proposes, and
  • 4. What you call "Americentric bias" is not, in fact, the system working as it should, per WP:RS, WP:N, etc..
CSB calls for increasing visibility of underrepresented areas, regions, and persons across Wikipedia. It doesn't call for a handful of impassioned users barring figures who are notable because of being notable in the U.S. I've noticed that you and Jim Michael 2 - obviously both favoring exclusion - comprise about 80% of the argumentation regarding the proposed inclusion of Barbara Walters, despite what seems to be a more even proportion of people arguing for her inclusion. It is, perhaps, a sign that certain areas of Wikipedia do need to be examined more closely to ensure that policy and guidelines are being appropriately applied, and that MOS:YEARS is in fact needed. 🌈WaltCip-(talk) 13:59, 1 February 2023 (UTC)[reply]
You seem to be implying that Scrubby & I have an anti-American bias. We don't; we're opposed to domestic figures being on main year articles because they're for international figures. When a well-known person from the Anglosphere (especially the US) dies, it's common for many fans to push for them to be added even if they're domestic figures. We shouldn't make exceptions on the basis of a deluge of fans demanding it. I'm pleased but very surprised that very few Anne Heche fans did so, because her death was one of the most reported of last year. I'm not going by who I know or am a fan of, because I've long been a fan of Robbie Coltrane but argued against him being included because he's a domestic figure.
It's unfortunate that this discussion, as well as those on WP:YEARS, are taking place when our best editor, Deb, is absent. Jim Michael 2 (talk) 18:00, 1 February 2023 (UTC)[reply]
Deb has been engaging in similar ownership behavior at WikiProject Years, but to a broader degree and for much, much longer. Thebiguglyalien (talk) 18:38, 1 February 2023 (UTC)[reply]
Exactly Jim; I believe those that are making such insinuations when we have both made ourselves absolutely clear about our positions and why we say there is systemic Americentric bias are simply ignoring what we have to say and painting false pictures of our attitudes and motives. How else can you jump to such ridiculous conclusions as to insinuate anti-American bias when all we are asking for at the end of the day is to end systemic bias and to treat figures/events from one country with the same standards as those from all other countries, and not have one standard for one country and another for everywhere else. Like, international notability isn’t a hard concept to grasp, and it’s insane that for some users here, such a notion as to base inclusion on the main international yearly pages on substantial notability beyond just one country/region and add domestically notable figures/events to the relevant Year In Country pages (even if we were to base these on “reliable sources” so as to demonstrate the subject’s international notability and impact) is unthinkable and unacceptable. Otherwise, as I said earlier here, to just completely ignore and/or intentionally twist and misinterpret what we have said and the concerns about the utter shambles of a system the Year pages ran under until mid-2021 (even those critical of aspects of the international notability criteria who would like to see it reformed such as InvadingInvader have conceded that these issues are valid and need to be addressed rather than swept under the carpet by conservative “status quo” users putting their fingers in their ears and repeating “reliable sources” over and over again) is not constructive in the slightest and at worst borders on bad faith accusations towards us. TheScrubby (talk) 23:16, 1 February 2023 (UTC)[reply]
You don't need to boldface giant portions of your text. I can read it just fine without it, thank you. WaltClipper -(talk) 12:46, 2 February 2023 (UTC)[reply]
  • I would like to echo what Cullen328 said above, noting that I've been an editor for almost 17 years and an admin for over 15, and have never seen the particular standard being bandied about in these discussions. The vagaries of "international notability" appears to have been created a short while ago by a small number of editors who see themselves as the sole gatekeepers of the year articles, and frankly, there is no clear guidance on how to establish who is, and who isn't, supposed to be listed in things like the "Birth" and "Death" lists in such articles. If I agree with one thing is that there should be some standard; insofar as listing everyone with a Wikipedia article who is either born in, or who has died in, a particular year is not reasonable, but insofar as we need to have such a site-wide standard, we need to have that discussion in a forum like this, and not in some backwater corner of Wikipedia, where some small group of editors makes a decision and decides to enforce it on everyone else that comes along later. That is decidedly not how we do things around here. We need objective standards that anyone who is unfamiliar with a person to be added to such a list can read and then assess a person against those standards. Those standards should be open to a broad discussion on a site-wide forum with wide participation. I'd be very much on board with whatever that discussion concludes. What I'm not on board with is some small cadre of editors coming up with their own rules, and then trying to enforce it on the rest of us. --Jayron32 15:58, 1 February 2023 (UTC)[reply]
  • FYI: Wikipedia:Village pump (proposals)#RFC: split births & deaths from year articles. Levivich (talk) 18:29, 1 February 2023 (UTC)[reply]
  • Noting that there is only one mention of the word "international" on WP:N, and it has nothing to do with "international notability". I echo Cullen and Jayron in the fact that this "international notability" criteria has no basis in any Wikipedia policy. We have just one standard for notability, not certain standards for certain pages made by WikiProjects that clearly decide to engage in ownership/gatekeeping contrary to Wikipedia's policies. JCW555 (talk)23:01, 1 February 2023 (UTC)[reply]
    So you believe what’s notable in one country/region is automatically notable internationally? TheScrubby (talk) 23:03, 1 February 2023 (UTC)[reply]
    Again, your "international notability" criteria has no basis in any Wikipedia policy. None. Zero. Zip. Ziltch. Nada. So stop using it as this shield you put up against any criticism of this arbitrary criteria that does not enjoy site-wide consensus. JCW555 (talk)23:12, 1 February 2023 (UTC)[reply]
    So you believe it is correct to fill main international yearly pages with domestic figures with scant international notability, and by extension promote systemic bias. I make absolutely zero apologies for believing that figures/events on the main international yearly pages ought to be internationally notable, and for believing that inclusion standards for figures/events from all countries should all be the same rather than arbitrarily having one standard for one/a handful of countries and another standard for everywhere else. TheScrubby (talk) 23:18, 1 February 2023 (UTC)[reply]
    There is only one standard we have, and that's Wikipedia:Notability. You'll notice that Wikipedia:InternationalNotability is a red link. Why? Because it's not any standard English Wikipedia uses. End of story. JCW555 (talk)23:29, 1 February 2023 (UTC)[reply]
    @TheScrubby, when editors talk about "notability", they mean "whether a given topic warrants its own article." They don't mean what the dictionary says. So when you say "He's not internationally notable", people are thinking "What do you mean? He's only getting a Wikipedia article in certain countries?"
    Wikipedia's inherently international. We either have an article that can be read by anyone in the world, or we don't have an article at all. There's no way to have a Wikipedia article that's international versus having a Wikipedia article that is only available in certain countries. All WP:Notable subjects are inherently "internationally" notable.
    I suspect that what you mean is that some subjects, though 100% WP:Notable for Wikipedia's purposes, do not have a public presence on the international stage.
    One possible way to deal with this is to adopt the Wikipedia:Vital articles approach. There's doubtless some reasonable number of desirable entries to include in each year. Whether that's 25 or 250, figure out the basic number. Then tell people: Make a list up to number±wiggle room. Maybe it's 50, and you want wiggle room of another 25. That would give each year 25 to 75 entries.
    But also tell them: Once we pass the magic threshold, if you want "your" entry included, you have to propose the removal of two existing entries, and those entries have to be similar to your proposed entry in some respect (e.g., you want to add sports, so you propose removing two sports entries; you want to add a politician, so you propose removing two politicians; you want to add an American, so you propose removing two Americans – the point is that people interested in a particular subject area can't propose removing subjects they don't care about). Editors then need to reach consensus about how to make space, before for your proposed new entry can be accepted. WhatamIdoing (talk) 23:52, 1 February 2023 (UTC)[reply]
    I think your comment is insightful, except adopt the Wikipedia:Vital articles approach is not an option for a mainspace article like year articles like 2022. Setting a numeric threshold is permissible for a WikiProject like WP:VA when including pages in its scope, but not for an article in mainspace. I don't think we'd want "Top 100 most important events of 2022" in mainspace. Levivich (talk) 00:09, 2 February 2023 (UTC)[reply]
  • I want to reïterate a point I made at the Walters RfC, and that Levivich has been getting at a bit above: Our judgment does not trump that of RS. Yes, at the margins, we may make editorial decisions like "Source X should be ignored because it's clearly sloppy reporting" and "Source Y seems to be factually incorrect here", but overall, our job as encyclopedists is to present information in due weight to what is represented in reliable sources. We do not get to say "I think the sources care about X, but I care about Y, so let's talk about Y instead". Telling, above, is the discussion about "The Slap" vs. the Peshawar attack. I'm certainly not going to sit here and argue that the former is objectively more notable than the latter, but, by what standard exactly is the converse being argued? Is the contention that it is intuitively obvious that a terrorist attack is more notable than an act of violence that was physically inconsequential but deeply culturally impactful? Many may feel one way or the other way, but that's the whole point of following reliable sources: It doesn't matter what we think.
    It troubles me deeply that some editors have not just appointed themselves arbiter of notability above the reliable sources, but indeed have explicitly argued that the reliable sources can't be trusted, to the extent that agreement with the reliable sources' perspective is disqualifying. Not to put too fine a point on it, but this goes against everything we stand for as Wikipedians. We follow the RS, even when we don't want to. As someone with some pretty out-there views, I regularly have to suck it up and write an article that follows statements in RS that I have personal objections to. None of us is above that.
    A stray thought in closing: Who says these articles need to be lists? -- Tamzin[cetacean needed] (she|they|xe) 06:02, 2 February 2023 (UTC)[reply]
    I just want to quickly address that stray thought. If anything, that's an even hotter subject of debate in the WikiProject. A few months ago, BorgQueen and I spent a few weeks expanding 2001 to bring it closer to the Good Article criteria (before and after), but a few regulars on the WikiProject shut it down and reverted all of the changes. Relevant discussions here and here for anyone interested. Thebiguglyalien (talk) 06:19, 2 February 2023 (UTC)[reply]
    Wow, this edit by InvadingInvader seems... Come on Tamzin, must not say V-word, must not say V-word... but I mean, seriously, that's removing 60kB of sourced material because "no consensus". Actually, not even "becase 'no consensus'", because no consensus yet. It's good-faith, so it's not vandalism, but it's functionally indistinguishable from it. -- Tamzin[cetacean needed] (she|they|xe) 12:04, 2 February 2023 (UTC)[reply]
    I will say that I believe InvadingInvader was trying to be a neutral party in the dispute. The larger problem is that several users on the WikiProject will challenge virtually every change or addition as "no consensus" without allowing discussion to take place. It's gotten quite out of hand, and it appears to have been going on since long before I got here. Thebiguglyalien (talk) 14:50, 2 February 2023 (UTC)[reply]
    It was more so procedural. At the time, I was not aware of the No Consensus essay that Tamzin cited. InvadingInvader (userpage, talk) 23:45, 2 February 2023 (UTC)[reply]
    While that edit was in good faith, this is a perfect illustration of the downsides of the "international notability criteria". The information that was removed was IMO due, informative, and encyclopedic. I largely align with Thebiguglyalien's views on prose and dueness when it comes to Years articles. Consistency among these lackluster articles seems to be valued more highly than quality. DFlhb (talk) 19:51, 3 February 2023 (UTC)[reply]
    In retrospect, it is. InvadingInvader (userpage, talk) 07:21, 6 February 2023 (UTC)[reply]
  • Reading above, the use of the term "Notability" seems to be causing a significant amount of miscommunication here, as that term has both a general meaning and a specific wikispeak meaning, which overlap, but also differ in important ways. (WhatamIdoing makes the point above as well.) Taking the premise that article should be somewhat in line with our guidelines on WP:SUMMARYSTYLE and WP:Article size, the question is not "notability", but (as Tamzin mentions) WP:DUE weight. There is a need to make sure cruft, and for that matter WP:RECENTISM, is handled appropriately, and no doubt these articles are a magnet for random trivia. However, while considering systematic bias is an important component of weight, it does not overcome the principle that articles should be based on reliable sources. We can evaluate those sources and consider how their target framework may or may not match ours, but that isn't the same thing as building purely based on our own evaluations. CMD (talk) 06:57, 2 February 2023 (UTC)[reply]
    WP:N is about being notable enough to have a WP article. All the events, births & deaths of each year are eligible to be in the relevant subarticles, such as 2023 in science & 2023 in the United States. Only those with substantial international notability are eligible to be in main year articles, such as 2023. Jim Michael 2 (talk) 11:13, 2 February 2023 (UTC)[reply]
    Again with the "international notability" criteria that has no basis in Wikipedia policy. JCW555 (talk)11:30, 2 February 2023 (UTC)[reply]
    Why do you think there are main year articles such as 2022 & country subarticles such as 2022 in China & 2022 in India? The main ones are international & the country articles are national. Jim Michael 2 (talk) 11:46, 2 February 2023 (UTC)[reply]
    The latter does not at all follow from the former. It makes sense that the main list will have a higher standard for inclusion than the national lists, but there's a leap of logic to there from the current "international notability" rules. What each list should be based on is what reliable sources say were the key events in that year in that region (or globally). In many cases, for the main list, this will coincide with "international notability", but not always. -- Tamzin[cetacean needed] (she|they|xe) 12:04, 2 February 2023 (UTC)[reply]
    This is all a very good example of why the term "international notability" should stop being used. Actual existing guidelines, such as WP:SUMMARYSTYLE and WP:DUE, are the relevant considerations. CMD (talk) 12:43, 2 February 2023 (UTC)[reply]
    It's logical & makes very good sense. You're saying that if some RS give the release of Squid Game more coverage than the end of the War in Afghanistan (2001–2021), we should do likewise, saying/implying that SG is more important?! Jim Michael 2 (talk) 15:44, 2 February 2023 (UTC)[reply]
    If some of them do? No. If the balance do? Yes. That's how being a neutral tertiary source works. -- Tamzin[cetacean needed] (she|they|xe) 22:36, 2 February 2023 (UTC)[reply]
    What benefit do we give to our readers from excluding widely-reported nation events but including the peaceful resolution of the Whiskey Islands dispute between Canada and Denmark? What benefit in the Deaths sectiondo we give to our readers by excluding Barbara Walters but including an Olympic gold medalist who most people have never heard of? I don't oppose the inclusion of both events/people, but it's baffling to trade those widely-reported events for niche ones only studied by academia. We're not Citizendium. InvadingInvader (userpage, talk) 19:32, 2 February 2023 (UTC)[reply]
    I've always been against mentioning the trivial dispute over tiny, barren Hans Island. A discussion about Olympians resulted in a consensus that individual Olympic gold medallists routinely be included. Jim Michael 2 (talk) 23:08, 2 February 2023 (UTC)[reply]
    Can you provide the link to this discussion? Thanks. —Bagumba (talk) 07:29, 6 February 2023 (UTC)[reply]

section break 2

Maybe the solution is to discontinue WP:YEARS & delete all Year related-pages. I suppose that'll never happen, but maybe it's an option to have in mind. GoodDay (talk) 22:18, 3 February 2023 (UTC)[reply]

Alternatively, someone stated above that turning the year articles into prose like how we do with 10th millennium BC, which could be a way to cut all the squabbling. Curbon7 (talk) 23:26, 3 February 2023 (UTC)[reply]
I mentioned above that I had done such edits to 2001 (before and after), but GoodDay and one other editor challenged the edits and forbade me from doing so any further because they felt I didn't get consensus from the WikiProject. Thebiguglyalien (talk) 23:46, 3 February 2023 (UTC)[reply]
Do you still wish to make those changes to those International Year pages? GoodDay (talk) 23:57, 3 February 2023 (UTC)[reply]
I think at this point the community needs to discuss more broadly what to do with the year articles moving forward, which is what I was trying to do when I opened this discussion. The years articles still follow basically the same format that they did at the beginning of Wikipedia (example), and they were never updated to reflect best practices like the rest of Wikipedia was. I think prose is a strong contender given the guideline at WP:PROSE, but whatever happens to them, it ultimately needs to be something that complies with modern Wikipedia standards, and I think it will take more than one or two users to figure out what that is. Thebiguglyalien (talk) 00:07, 4 February 2023 (UTC)[reply]
Would your proposal bring about the deletion of the image collages? I'm asking, as those additions were the result of a compromise. GoodDay (talk) 00:16, 4 February 2023 (UTC)[reply]
Personally, I'm not a huge fan of the image collages. But again, it really comes down to what the community thinks, and broader issues of structure and coverage will probably take precedence over cosmetic concerns, at least until this mess is cleaned up a little bit. Thebiguglyalien (talk) 00:26, 4 February 2023 (UTC)[reply]
Years are clearly an encyclopedic topic. How could deleting all year articles be a solution? If you mean to avoid disagreements, the same could be said for many other types of articles, including those about politics, religion, sex, sport, entertainment, celebrities etc. Jim Michael 2 (talk) 15:15, 4 February 2023 (UTC)[reply]
  • To the main point of this discussion: I do not believe that the creation of an MoS subpage (specific to year lists) is warranted (on neither the basis of this discussion nor the previous discussions giving its rise). While disagreement (legitimate and sincere on both sides) clearly exists, the path of its mitigation is not within the style manual's remit proper. That path (hard traveled and often obscure) is through editorial consensus as reflected in the letter and spirit of Wikipedia's policy/guideline provisions. Provisions, of which, the MoS itself is subject and subordinate to. There's no magic pill to cure the ills of hard work needed to resolve editorial disputes like this, yet the pain and scar tissue of the hard fight can easily be magnified by needless expenditures of time and effort from asking the wrong question or pursuing its answer in the wrong place. It's clear that general notability guidelines (as written) can not be objectively used to segregate a topic by imagined increments of notability from regional to global. It's not clear whether the community desires such a means or not (this should be known first). And then, if it does, an RfC capable of amending our notability guidelines would have to ensue. Use the time available before the RfC to establish a well written request that addresses as many concerns as possible and effectively achieves the stated goals. For example, I do not believe efforts to establish "international notability" will satisfactorily identify a topic as being globally significant, if that is your goal, because international can easily be limited to a geographical region like Europe, The Middle East, Africa, and others that may not be sufficient enough to satisfy global importance, if that is your ultimate goal. These things said, I believe the editorial dispute outlined in this discussion is intractable, long standing, and worthy of being resolved; I wish you the best and look forward to what the future may bring. Sincerely. --John Cline (talk) 00:56, 5 February 2023 (UTC)[reply]

Expanding WP:NOTGUIDE to include part lists/specifications

There are currently many hundreds of articles on Wikipedia which are just lists of information like gear ratios in every variant of every car part ever made by every single manufacturer. Pretty much every single page in every one of these categories is huge amounts of non-contextual information without evidence of notability. Here are some random examples:

I'm not sure what the solution is here. A while ago I opened Wikipedia:Articles for deletion/Toyota C transmission after stumbling upon the article from random and the consensus there was to delete, but it just feels painful. Following this rough consensus, the follow-up should be to go on a mass nomination spree and start spraying WP:TNT since I don't think there's much hope any of these articles meet GNG or WP:NPRODUCT. On the other hand, running around as a deletion monkey and blowing up megabytes of almost certainly reliable text that someone painstakingly copied from a manual doesn't feel good either.

WP:NOTGUIDE is much less clear here than it could be. It's clearly the most relevant policy alongside WP:NOTDATABASE and WP:NPRODUCT, but only explicitly condemns "how-to" style guides which these are not. I'm unsure about the right way to proceed. I feel like a modification to NOTGUIDE might be due here to clarify explicitly whether "parts list" style articles are appropriate. BrigadierG (talk) 00:40, 3 February 2023 (UTC)[reply]

For a proposal to mass delete articles that are all under a particular wikiproject, good etiquette would say to at least notify that project. Otherwise it looks like you are trying to get the deletions done before the involved editors know what's happening. I came across this topic by sheer luck. I have notified them at Wikipedia_talk:WikiProject_Automobiles#Proposal_to_delete_many_engine_and_transmission_articles.
The C transmission article got consensus to delete mostly because of too few references rather than whether it was a bad choice of article.  Stepho  talk  00:36, 4 February 2023 (UTC)[reply]
I'm not trying to propose mass deletion. If I wanted mass deletion I wouldn't waste my time talking about it here, I'd just start mass nominating, and I'm sure nominating enough would successfully ram a good few through successfully given the tenuous connection these articles have to GNG. Deletion isn't necessarily my goal. What I'm seeking is clarity, and perhaps an alternate route forward. I'm not sure the current article structure is appropriate, but it might be possible to do something like folding down the amount of information (which I think is generally excessive) and grouping articles as series rather than as individual articles for each part. As for "too few references rather than whether it was a bad choice of article", if you can find 2 quality secondary sources to pass WP:NPRODUCT then I will personally recreate the article. BrigadierG (talk) 02:40, 4 February 2023 (UTC)[reply]
I can see the problem with these articles – they lack sufficient context and fail to explain what they are about. The question is whether they provide any useful information to the average reader. Considering that most people seem to be interested primarily in tech specs, one could argue that this is the case, but considering the poor referencing situation I'd personally say that we sadly got nothing but a hearsay compilation. Especially when browsing through the history of these articles I start getting serious doubts about the accuracy of the information provided. And I guess that no article is always better than an article that is made up of unverifiable nonsense. Best regards, --Johannes (Talk) (Contribs) (Articles) 08:22, 4 February 2023 (UTC)[reply]
I originally nominated the article after an IP changed an unreferenced gear ratio to add a "XD" Toyota gearbox specification to a long list of gearbox specifications. Their only other edit was unconstructive so I reverted it, but I really have no idea either way, and I didn't feel like it was appropriate to have to trawl through a 200 page specification to find out. Call it inadequate referencing or whatever, but a nightmare from a WP:V perspective BrigadierG (talk) 00:49, 5 February 2023 (UTC)[reply]
The problem is in fact wider: restaurants with their menus, record labels with all their signings (notable or non-notable). The Banner talk 16:21, 4 February 2023 (UTC)[reply]
Support clarification. Those articles also fail INDISCRIMINATE and NOTDIRECTORY and our basic standards for DUE and fancruft and should be deleted. JoelleJay (talk) 20:52, 4 February 2023 (UTC)[reply]

Plot summary structure

Looking at the plot summaries for films on Wikipedia, it seems like in general they tend to devote a disproportionate amount of words summarising the second half of the film, and ignore many plot points from the first half. (I haven't looked at the books much, but it's possible they have the same problem.) I wonder if there may be some kind of recency effect going on, in that editors remember the second half of the film much better because they saw it second. Regardless, is there anything that can be done about this? Plokmijnuhby (talk) 16:02, 6 February 2023 (UTC)[reply]

How about excluding plot summaries entirely, and basing article content on what reliable sources have to say about the subject instead? AndyTheGrump (talk) 16:06, 6 February 2023 (UTC)[reply]
As plot summaries are to be concise and capture that main points of a film without introducing interpretation, it is more than natural to have these focus on a film's resolution than introduction. Masem (t) 16:09, 6 February 2023 (UTC)[reply]
This happens due to a bad rule that we have, called WP:PLOTSOURCE: The plot summary for a work, on a page about that work, does not need to be sourced with in-line citations, as it is generally assumed that the work itself is the primary source for the plot summary. This is one of the few areas of Wikipedia where--for reasons I can guess--we have a guideline that violates a policy (WP:NOR). Works of fiction are primary sources for their own plots, and normally we don't allow editors to interpret primary sources because that's original research. Notable works of fiction has WP:RSes about them, of course (reviews), and those include plot summaries. What we should do is require all plot summaries in Wikipedia articles to be summaries of plot summaries written by RSes. That way, we can be sure that it is the RSes, and not editors, who are interpreting the plot. "Interpreting the plot" means summarizing it, identifying important plot points, conveying how they are related, etc., all of which should be done by RSes. It's not a surprise, then, that when editors summarize a plot by themselves (instead of summarizing RS plot summaries), they end up summarizing the plot wrong. Levivich (talk) 16:22, 6 February 2023 (UTC)[reply]
I agree, but the headwind on this is very strong. Donald Albury 16:28, 6 February 2023 (UTC)[reply]

RFC: Clarifications to WP:COSMETICBOT for fixing deprecated HTML tags

This is a request for comment regarding Wikipedia:Bot policy, specifically the section on restricted bot tasks related to cosmetic changes outlined at WP:COSMETICBOT. Ultimately, I would like to start a discussion about whether or not the bot account User:MalnadachBot is operating within Wikipedia policy, and whether WP:COSMETICBOT needs to be updated to clarify how these kinds of bots should be handled in the future. First, a bit of history behind this situation:

MalnadachBot (talk · contribs) is a bot account that is owned by User:ಮಲ್ನಾಡಾಚ್ ಕೊಂಕ್ಣೊ. This is an AWB bot that primarily works on fixing lint errors, which are common technical issues in wikitext markup. A list of lint errors detected on Wikipedia is available at Special:LintErrors. The bot has had a total of 13 tasks approved by the Bot Approvals Group. The first 11 tasks were fairly standard error-fixing tasks, each of which had a limited scope to fixing a specific type of error. Task 12 was requested to fix another specific error associated with deprecated <tt>...</tt> HTML tags. However, the BAG member who approved the task (in this case, User:Primefac) decided to expand the scope of the task to fixing all linter errors across Wikipedia, and speedily approved the task. Since that task approval, MalnadachBot has made millions of edits, becoming one of the top accounts by edit count in Wikipedia history.

One of the most common edits that the bot makes is to replace <font color="red">...</font> tags with <span style="color:red;">...</span> tags. This is because <font> tags were technically deprecated when HTML5 was introduced, so current best practices for writing new HTML are to use span tags instead of font tags for customizing the font and style of text. If you inspect the bot's contributions, you'll notice that the vast majority of its edits are in the Wikipedia, Wikipedia talk, User talk, and Article talk namespaces. This is because using custom font styles is relatively uncommon in WP articles themselves. However, they are very common in many user's signatures (including my own). So, MalnadachBot is essentially going through nearly every discussion that has ever taken place in the 20+ year history of Wikipedia, and updating old user signatures to use proper HTML. Many of these discussion pages have not been edited in 10+ years, and many of them receive nearly zero annual pageviews. It has made millions of such edits, and there are many millions more in the queue to be done. Over the past 12 months that this task has been running, many users have complained on the bot owner's talk page and Wikipedia:Bots/Noticeboard that the bot is clogging up watchlists and revision histories with millions of trivial, purely cosmetic edits, but there has been no wider community discussion to come to a consensus on how to move forward. If you used to use <font> tags in your signature, you've no doubt had your watchlist bombed by MalnadachBot in the last year.

One thing to note is that these kinds of edits do not result in a change to the way that these pages are rendered to the user. This is because, despite <font> tags being deprecated in HTML5, all major web browsers still understand and support the older deprecated tags, and no major browser has announced an intention to stop supporting these tags in the foreseeable future (which is why these specific errors are categorized as the lowest priority on Special:LintErrors). Therefore, these edits are purely cosmetic changes to the wikitext of many pages, that do not change the visual rendering of the page. Typically, these kinds of edits are not allowed to be done on their own by bots (per WP:COSMETICBOT), and can only be done if they are combined with other substantive edits that do cause a change to the visual rendering of the page. However, there is one exception noted within WP:COSMETICBOT for fixing "egregiously invalid HTML such as unclosed tags, even if it does not affect browsers' display or is fixed before output by RemexHtml (e.g. changing <sup>...</sub> to <sup>...</sup>)". It has been asserted that MalnadachBot's task 12 falls under this exception, but the wording of the policy leaves much room for different interpretations.

So, the primary question I'd ask for your input on in this RfC is: does <font color="red">...</font> constitute "egregiously invalid HTML" that should be proactively fixed by a bot, or should it be considered a purely cosmetic change to the wikitext that should not be performed by a bot?

And the secondary question in this RfC is surrounding whether or not the last bullet point in WP:COSMETICBOT should be expanded to be more specific about defining exactly what types of errors are considered "egregious", and which aren't. A proposal for an updated wording can be found below.

—⁠ScottyWong⁠— 08:33, 7 February 2023 (UTC)[reply]

Primary question: are deprecated HTML tags considered egregiously invalid?

Please indicate below if you believe that automated fixing of deprecated HTML tags should be considered "egregiously invalid HTML" and therefore be permissible per the current wording of WP:COSMETICBOT, even if those deprecated HTML tags are still supported by all major browsers. (Support will be interpreted as support for the bot task to continue, Oppose will be interpreted as agreement that the bot task is in violation of policy and should be halted.)

  • Oppose as RfC author. I don't believe that fixing deprecated but supported HTML tags should be considered "egregiously invalid HTML", considering that every major browser still supports them, and will continue to support them for the foreseeable future. Fixing user signatures on ancient discussion pages is a purely cosmetic change to wikitext that does not alter the visual rendering of the page. This bot task is only generating millions of trivial edits that clog up user's watchlists and revision histories. While it's possible to ignore these bot edits, the bigger question to be discussed here is whether or not those edits should be happening at all. —⁠ScottyWong⁠— 08:33, 7 February 2023 (UTC)[reply]


Secondary question: should WP:COSMETICBOT be updated?

Please indicate below if you believe that WP:COSMETICBOT should be updated to be more specific about automated fixing of deprecated HTML tags that are supported by all major browsers. I will propose an updated wording, but please feel free to propose your own if you like. My proposed rewording is: "egregiously invalid HTML such as unclosed tags, even if it does not affect browsers' display or is fixed before output by RemexHtml (e.g. changing <sup>...</sub> to <sup>...</sup>). This does not include fixing properly formed HTML tags that have been deprecated, as long as the tags are still supported by all major web browsers."