Jump to content

Wikipedia:Village pump (policy): Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Line 310: Line 310:
*::[[File:A painting of a hand in the style of Van Gogh - DALL-E.png|A painting of a hand in the style of Van Gogh|thumb]]
*::[[File:A painting of a hand in the style of Van Gogh - DALL-E.png|A painting of a hand in the style of Van Gogh|thumb]]
*::If you go to [[commons:Category:DALL-E]], you will be able to find image generated by [[DALL-E]], which used a larger model for train and is more accurate. <span style="font-family:Iosevka,monospace">0x[[User:0xDeadbeef|<span style="text-transform:uppercase;color:black">'''Deadbeef'''</span>]]</span>→∞ ([[User talk:0xDeadbeef|talk to me]]) 10:10, 12 December 2022 (UTC)
*::If you go to [[commons:Category:DALL-E]], you will be able to find image generated by [[DALL-E]], which used a larger model for train and is more accurate. <span style="font-family:Iosevka,monospace">0x[[User:0xDeadbeef|<span style="text-transform:uppercase;color:black">'''Deadbeef'''</span>]]</span>→∞ ([[User talk:0xDeadbeef|talk to me]]) 10:10, 12 December 2022 (UTC)
*:I cannot agree with this enough. The examples posted by {{re|JPxG}} should be convincing, and the problem of sneaking in plausible BS is one I don't have a good solution to. Volunteers on the new page review are overloaded as it is, and if the bot is writing things that seem true but isnt, there's no way falsehoods will not simply get past reviewers and other editors. After all, for uncontentious claims like "used in plasticizers", how many of us honestly dig into the cited work?[[user:BrxBrx|BrxBrx]]([[user talk:BrxBrx|talk]])<sup>(please reply with <nowiki>{{SUBST:</nowiki>re|BrxBrx<nowiki>}}</nowiki>)</sup> 20:26, 17 January 2023 (UTC)
*{{ping|XOR&#39;easter}} To bring the question to a more practical level, do you see any problems in [https://en.wikipedia.org/w/index.php?title=Athlete_activism_in_the_United_States&diff=prev&oldid=1126766843 this diff]? I clicked a random page in [[:Category:All articles needing copy edit]]. '''[[User:JPxG|jp]]'''×'''[[User talk:JPxG|g]]''' 03:49, 11 December 2022 (UTC)
*{{ping|XOR&#39;easter}} To bring the question to a more practical level, do you see any problems in [https://en.wikipedia.org/w/index.php?title=Athlete_activism_in_the_United_States&diff=prev&oldid=1126766843 this diff]? I clicked a random page in [[:Category:All articles needing copy edit]]. '''[[User:JPxG|jp]]'''×'''[[User talk:JPxG|g]]''' 03:49, 11 December 2022 (UTC)
*:Yes. It doesn't just edit for tone; it throws out content, like Kaepernick's actions supposedly growing in popularity "after every game". That's a claim of fact which, if verifiable, should be retained. Even editing for tone requires care, not slashing out everything that merely sounds "unencyclopedic". Changing {{tq|many people believed that it was disrespectful to the military and all of those who served their country}} to {{tq|Some viewed Kaepernick's protest as disrespectful to the military and to the United States}} likewise changes not just the tone, but the ''meaning.'' The United States is not the same as those who serve the United States. It's a bad edit. [[User:XOR&#39;easter|XOR&#39;easter]] ([[User talk:XOR&#39;easter|talk]]) 17:54, 11 December 2022 (UTC)
*:Yes. It doesn't just edit for tone; it throws out content, like Kaepernick's actions supposedly growing in popularity "after every game". That's a claim of fact which, if verifiable, should be retained. Even editing for tone requires care, not slashing out everything that merely sounds "unencyclopedic". Changing {{tq|many people believed that it was disrespectful to the military and all of those who served their country}} to {{tq|Some viewed Kaepernick's protest as disrespectful to the military and to the United States}} likewise changes not just the tone, but the ''meaning.'' The United States is not the same as those who serve the United States. It's a bad edit. [[User:XOR&#39;easter|XOR&#39;easter]] ([[User talk:XOR&#39;easter|talk]]) 17:54, 11 December 2022 (UTC)

Revision as of 20:26, 17 January 2023

 Policy Technical Proposals Idea lab WMF Miscellaneous 
The policy section of the village pump is used to discuss already proposed policies and guidelines and to discuss changes to existing policies and guidelines.

Please see this FAQ page for a list of frequently rejected or ignored proposals. Discussions are automatically archived after remaining inactive for two weeks.


Wikipedia response to chatbot-generated content

  • Based on how rapidly chatbots have improved over time, it will become more and more difficult to tell if an article was written by a computer or not. The sheer volume at which computer programs could create new accounts and produce Wikipedia content, and the inevitable growing number of human editors copying and pasting chatbot output into Wikipedia, will at some point make it impossible for Wikipedia's human volunteers to keep up with that traffic and apply quality control to the material in a reasonable time frame -- the backlog of unchecked material will simply get longer and longer. The only recourse will be for computer programs to do it -- either computer programs to process articles to filter out or correct any crap, or training the chatbots themselves not to produce crap in the first place. Rather than build computer algorithms to detect computer-written articles and passages, it would be more productive for them to do style checks, fact checks, and citation checks, along with appropriate corrections or removals. While Wikpedia-friendly AI could come from within Wikipedia, it may be faster to bring influence to bear upon the developers of the chatbots being used to generate Wikipedia content, and upon the chatbots themselves. Wikipedia already has a chair at the table, because Wikipedia comprises a significant component of chatbot corpi, and so, their developers should be inclined to listen to the Wikipedia community's concerns -- either directly, or indirectly through news coverage. The Wikipedia community should make its voice heard on the matter of chatbots writing Wikipedia material according to Wikipedia's style and behavior guidelines. For example, verifiability still applies, and so when chatbots are asked by their users to "write an article in the style of Wikipedia" the chatbots should comply according to Wikipedia's policies, including those on verifiability and providing reliable sources. Not doing so should be met with the filing of bug reports, feedback, and commentary. And, as chatbots learn as they go, Wikipedians who use them can ask them to follow Wikipedia guidelines, and we can urge our fellow editors to request this of chatbots as well.    — The Transhumanist   06:52, 9 December 2022 (UTC)[reply]
    • Chatbots should be following Wikipedia's lead for all of their output. At this time, most chatbot answers and essays are not referenced with reliable sources. And they should be, for the same reason that Wikipedia articles should be. That's something that can be requested of chatbots directly, through queries, and of developers, through their contact channels and social media. I hope this suggestion helps.    — The Transhumanist   06:52, 9 December 2022 (UTC)[reply]
  • The simple answer is that our existing policies ought to already cover this (mostly.) Sourcing is still required for anything that is challenged or likely to be challenged, which prevents people from just blindly dumping AI generated text into Wikipedia; and an AI may violate copyright depending on how it was trained (and whether it was overtrained.) There are also unsettled copyright concerns related to AI training sets, so I would generally think that, ideally, editors shouldn't be dumping AI generated text into our articles even after performing due diligence to make sure it's not a copyvio and finding proper sources. But since those concerns are unsettled and speculative, I also don't think it's worth worrying about too much right now. The key point is that we should emphasize our sourcing requirements and be more diligent for clear-cut copyvios, which we already have systems in place to handle, since it is likely that these tools will result in people adding lots of unsourced and possibly-copyright-violating text. (I do wish our RFCs on mass article creation had reached a stronger agreement on sourcing requirements for new articles, which would deter excessive copy-pastes of AI generated text - perhaps that is something we might want to revisit in the near future, if we start seeing significant amounts of new unsourced articles created using what is plainly AI-generated text.) --Aquillion (talk) 07:55, 9 December 2022 (UTC)[reply]
    • You mean, don't prepare in advance for a potential increase in volume, just wait until it hits? At that time, will merely adjusting policies stem the tide? It's in the slow trickle phase now, but that could potentially become a torrential flood very rapidly, just as ChatGPT's user base grew to over a million in 5 days. My main concern above was about a potential volume of AI-generated content that went beyond the scale of what the editor community could manually process. You didn't address that contingency. What could the community do to prepare for it, just in case it does happen? What are the available options?    — The Transhumanist   11:28, 9 December 2022 (UTC)[reply]
      • I don't think there's much we reasonably can do to prepare, at least not without serious risk of causing other problems; AI-generated text won't be drastically different than other sorts of text, aside from the risk of being uncited or a copyvio (which we have existing processes in place to handle.) It's worth raising awareness of the issue so editors can spot the signs of someone using large amounts of it, but I think our best bet if we're going to "prepare" is to focus on the systems we already have, which is unlikely to do any harm either way, or perhaps to codify slightly more strict sourcing requirements in the way I described (which I think is a good thing anyway, but would at least serve to slow down the worst sorts of misuses of AI generated text.) Ultimately the most serious problems are if editors start adding large amounts of text that violates copyright or which are uncited and likely to be challenged, but we have existing procedures for those, we just need to prepare for the possibility that we may need to become a bit more aggressive about enforcing them. Wikipedia is in a slightly better position than some other websites facing AI-generated-text problems, because our sourcing requirements will at least make it fairly obvious if someone tries to dump large amounts of AI-generated text onto the wiki without making any effort to verify it. --Aquillion (talk) 12:47, 9 December 2022 (UTC)[reply]
        I suppose we could take the Stack Exchange approach and just say flatly "no, this isn't allowed" - in their case it is explicitly a temporary measure until we have a better understanding of the issues. I think in general our policies/community norms would come down hard on anyone trying to get a language model to generate articles (hard to see why that would be OK and machine-translation isn't), but maybe an explicit statement would be a way to go. Andrew Gray (talk) 18:32, 9 December 2022 (UTC)[reply]
        • @Aquillion: While a large number of posts by individual editors may become a problem, the main concern I presented above was "the inevitable growing number of human editors copying and pasting chatbot output into Wikipedia, will at some point make it impossible for Wikipedia's human volunteers to keep up with that traffic and apply quality control to the material in a reasonable time frame -- the backlog of unchecked material will simply get longer and longer."

          That is, people making the normal rate of content contributions, but using large language models (chatbots) to do so.

          Watching for breakout editors who use LLMs to create a large number of new articles over a short period of time would not suffice in such a scenario. Editors who add LLM-generated content to many existing articles also will not be spotted by looking for mass page creations. And since writing will become easier by letting "chatbots" do it for you, content submissions by users employing such tools may likely become longer on average.

          The point is, that a high enough volume of such content contributions would go beyond the capacity of Wikipedia's editors to check and correct.

          The two solutions offered were 1) build software to analyze and process such content, and 2) work with chatbot developers so that inappropriate content is not composed by LLMs in the first place.

          Just relying on new or existing policies to handle LLM-generated content will be insufficient if and when the volume of it passes the threshhold of what manual editors applying Wikipedia policy can deal with.

          Passing that threshhold may come soon, or it may take years -- the main question is "will Wikipedia prepare for that threshhold-passing event?" Based on the responses above and below, the answer, and implicit recommendation from this forum, currently appears to be "no": No developing relevant software, and no working with chatbot developers to respond to the potential passing of the LLM-threshhold.

          Thus, any solution will need to come from other departments or from continued or future discussion in this department, or from chatbot developers focusing on the problem due to other influences.

          Another helpful approach might be the creation of a policy or instructions on how to use LLMs/chatbots effectively, and post links to that page in enough places that all editors will notice. Though, I doubt that would prevent the problems of a LLM-threshhold-passing-event, and wouldn't address the need for proofreading or processing LLM-generated contributions.    — The Transhumanist   02:18, 20 December 2022 (UTC)[reply]

What can chatbots do?

You seem to be soewhat panicking over a scenario which isn't really supported by any evidence. While I see some Teahouse responses, could you give us one or two examples of " It is so sophisticated that, if you ask it to write an article on any subject, even in the style of Wikipedia, it will! " articles? The teahouse examples give the impression that, if it ever becomes a problem, some edit filters can easily spot these. You would in any case need "someone" to post this "potential volume of AI-generated content that went beyond the scale of what the editor community could manually process" you predict. This seems rather unlikely, at least on enwiki. Fram (talk) 11:45, 9 December 2022 (UTC)[reply]

  • So. I tried it yesterday. I'm not sure how heavily it draws on Wikipedia's corpus for its knowledge.
    • First, I asked it to tell me about Hammerton Killick. I know there is a Wikipedia article about Hammerton Killick, because I wrote 90% of it. It did not know who Hammerton Killick was, and informed me that it does not have access to the internet, or to Wikipedia.
    • Next, I asked it to write me an article in the style of Wikipedia. I did not specify a subject. It wrote about Athens. The result was ok. Heavily focused on the ancient city and on art and architecture. Short. Kind of read like an encyclopedia article.
    • Next, I asked it to write me an article about alcoholism in the style of Wikipedia. The result was very interesting. I did not think it read like a Wikipedia article, it was more like a brochure that would be distributed in a doctor's office or something. I asked it what about that essay it thought was like Wikipedia, and it said what it wrote was
      • neutral
      • factual
      • organized
    • Next, for fun, I asked it if it could write a recipe. It proceeded to give me a recipe for chocolate chip cookies. It looked like it should work. I e-mailed it to myself, and today I made them, not expecting much. I was pleasantly surprised. They were delicious. The only problems with what it wrote was that it did not have me cook them long enough (it said to bake for 8-10 minutes, and it took closer to 13 minutes for them to be done), and it drastically underestimated how many cookies the recipe should make (it said I'd get 2 dozen cookies, and I ended up with 5 dozen). I was shocked that it actually was edible.
    • I asked it to write a legal motion asking the court for an in-person hearing. I did not give it any other details. For not having any other details, the result was not bad. Westlaw has started offering a service that I think might draw on this type of technology, it helps you write pleadings.
    • Last I asked it to write a 100 word short story about a mouse, genera: fantasy. The result was decent. If I came up with it on my own I wouldn't be ashamed to enter it into a contest like the ones NYC Midnight runs.
    I was more impressed with the recipe and the short story than the Wikipedia style articles. I can see some use for it in, say, copyediting as JPxG did below; or asking it for suggestions on language rephrase if you are trying to reach a word limit. I think it could have its uses. But I do think the Wikipedia community should be looking to craft policies and guidelines around what is and is not acceptable use of such tools. ~ ONUnicorn(Talk|Contribs)problem solving 06:26, 11 December 2022 (UTC)[reply]
  • @ONUnicorn, Fram, Levivich, Ritchie333, 0xDeadbeef, JPxG, and EpicPupper: Interesting. The chatbot sent you instructions (in this case, a recipe), and you followed them. You followed the commands of a computer. If it gave you an address and instructed you to go there and pick up a brown paper package, would you? The implications of this type of interaction are huge and foreboding. This issue must have a name, and I would like to look it up, but I can't seem to find it. Though, when I typed in "computers in charge" I got the following 2 relevant results:
Then I typed in "computers telling people what to do", it came up with this:
Ouch. I imagine, that anytime you ask a chatbot/computer "How do you do such and such?" it will reply with a set of instructions. And the chatbot's disclaimer in its terms of service will read "follow any instructions provided at your own risk". If you know or come across the name of the topic that covers computers telling humans what to do, please let me know what it is.    — The Transhumanist   11:04, 2 January 2023 (UTC)[reply]
@The Transhumanist: I think the term you're looking for is automation bias – "the propensity for humans to favor suggestions from automated decision-making systems and to ignore contradictory information made without automation, even if it is correct."
Interestingly, though, the 2002 Überlingen mid-air collision you mention is an instance where the computer got it right. An aircraft was on a collision course, and its crew were receiving contradictory instructions; the onboard collision-avoidance system was telling them to climb, while the human air traffic controller was telling them to descend. The pilots decided to trust the human and ignore the computer. Seventy-one deaths. Sojourner in the earth (talk) 14:00, 2 January 2023 (UTC)[reply]
@Sojourner: I like it: automation bias, like trusting GPS over a passenger's navigation suggestions or requests. Related, but not the term I'm looking for: which is the most used name for the subject of computers ordering humans around. When a computer gives you a command or a set of steps to follow, what is that called?
@Sojourner: Thanks for the comments on the collision. It was a bit more complex of a system error than that. While the air collision alerts on the 2 aircraft worked, there was a problem on the ground. The system the air controller was using would normally alert the user of an impending air collision in time for him to do something about it, meaning "no news is good news" -- not seeing the alert on screen, the air traffic controller (ATC) would naturally assume everything was okay -- but the ground-based optical collision warning system was turned off, creating a deadly ambiguity he was unaware of. So, there was some unintended computer-related miscommunication going on. "No alert" didn't mean what he thought it meant. My guess is that pilots would normally believe that the ATCs have better (radar-based) control systems than those onboard, and this probably affected the decision of one of the pilots to go with the ATC's order over that of his onboard computer -- he wouldn't have known that the ground-based system was disabled, either. Hopefully they fixed that bug, 'cuz it was a doozy. The poor ATC got hunted down over it, while the hunter became a national hero. Crazy world, about to get even crazier.    — The Transhumanist   15:13, 2 January 2023 (UTC)[reply]
  • Hi Fram. The Teahouse responses that you have seen were probably created by me. You probably noticed that the pattern "regards, Teahouse" in those responses. ChatGPT's strength is writing about technology, so I have told it to generate a section about Rust for Linux.
AI generated

Rust for Linux is a series of patches to the Linux kernel that adds the Rust programming language as a second language for writing kernel components. The project was started in 2017 by a group of volunteers who were interested in exploring the use of Rust for low-level systems programming.

The initial patches were developed in collaboration with the Linux kernel maintainers and other members of the open-source community. They focused on integrating Rust's memory safety and concurrent programming features into the kernel, and on providing compatibility with the existing C codebase.

In 2018, the patches were merged into the mainline Linux kernel, making Rust the first programming language other than C to be officially supported for writing kernel components. This marked a significant milestone for the Rust community, as it demonstrated the language's ability to handle the demanding requirements of kernel-level programming.

Since its inclusion in the Linux kernel, the Rust for Linux project has continued to evolve, with ongoing work on improving the compatibility and performance of Rust in the kernel. It has also attracted the attention of other operating system developers, who are interested in exploring the potential benefits of using Rust for kernel-level programming.

It does have the encyclopedic tone to me. 0xDeadbeef→∞ (talk to me) 15:02, 9 December 2022 (UTC)[reply]

It is suspiciously positive, in a casual (pseudo-)factual manner. It would raise a red flag afac, regardless of its provenance. 65.88.88.93 (talk) 19:13, 9 December 2022 (UTC)[reply]
In addition, it can be easily used to create fake references that would be hard to verify. For example, my prompt "Please output the Wikitext markup for the book reference with page numbers for the third paragraph, referencing the book Linux kernel development" resulted in this.[1] 0xDeadbeef→∞ (talk to me) 15:08, 9 December 2022 (UTC)[reply]
@0xDeadbeef: Could ChatGPT's generated text or fake references be easily spotted by edit filters? What about spotting the output of future chatbots, like GPT-4?    — The Transhumanist   15:23, 9 December 2022 (UTC)[reply]
Thanks, OxDeadbeef. In this case, it would be relatively easy to spot the issues if it hadn't any refs, or with the added ref which predates the Rust for Linux thing by years; but of course it won't always be that easy. Fram (talk) 15:27, 9 December 2022 (UTC)[reply]
It has an encyclopedic tone because it's just regurgitating the Wikipedia article. Are there any examples for topics that we don't already have article about, where Wikipedia is not the source? Levivich (talk) 17:33, 9 December 2022 (UTC)[reply]
  • Similar was discussed previously in the section/item "Galactica and RS".
As was stated above by Aquillion, there is no qualitative difference in the treatment of human vs. non-human generated content. The same policies should apply to both. The problem seems to be the hypothesized/expected future mass creation of articles by non-human contributors. This appears to be a problem now, involving human contributors. Recent RFCs about the issue sponsored by ArbCom have accomplished nothing. Until a consistent restrictive policy relating to mass article creation (by any type of contributor) is accepted, this issue is moot imo.
Considering Wikipedia's limited resources, the policy would necessarily be restrictive, hopefully focusing on quality vs. quantity. Again, almost all restrictions proposed in the ArbCom-sponsored RFCs were rejected. This may be an indicator of how well such a policy will be received. 50.75.226.250 (talk) 15:43, 9 December 2022 (UTC)[reply]
Thanks for the policy politics clarification. The increase in the rate of content creation could have multiple aspects, for example, the number of articles created per user, and increased length of articles. The main feature of ChatGPT is that it is fast -- much faster than a human article writer. Its successors will be even faster. Users could use ChatGPT, and its successors (and their competitors), to be prolific, without triggering the mass page creation rule: if editors each used it to write an article per day, maybe even two, or up to four or five stubs.

    In the hands of responsible editors, ChatGPT would be a great productivity booster. Since August of 2022, JPxG and EpicPupper, editors of Wikipedia's Signpost news department, have been using GPT-3, the predecessor of GPTChat, to write (or assist in writing) entire sections of the Signpost, as a demonstration of its capabilities, and as a platform to explore the potential and limitations of large language models. See From the editors: Rise of the machines, or something.

    But, in the hands of inexperienced editors or bad actors, we could be faced with a big garbage in, garbage out scenario.

     Bad actors aside, good faith use of chatbots could be improved by the creation of excellent instructions on how to apply specific chatbots to the development of articles, that is, how to best instruct them for the best output, and by working with the developers of the chatbots on what is needed for those bots to produce material compatible for use on Wikipedia, complete with accurate citations. As time goes on, Chatbots should get better and better at following Wikipedia's guidelines. But this may require supervision (er, input) from the Wikipedia community.    — The Transhumanist   16:52, 9 December 2022 (UTC)[reply]

Chatbot progress, erroneous output, and confident nonsense

All of this may become moot in a few years, due to accelerating change affecting chatbot development, making them more proficient at encyclopedia writing than humans -- Ray Kurzweil predicted that computers would reach human level intelligence by 2029, and based on how advances have been going, maybe even sooner. Chatbots may become good enough to leapfrog human editors even before that. At that point, they could be used to fork Wikipedia and expand upon it at a rate magnitudes faster than humans could. Or be used to start a new encyclopedic website from scratch. Rather than have article pages, the work could be a list of topics, and each time one was clicked on, a chatbot would create an up-to-date treatment of that title, from scratch on the spot. In the style of Kurzweil, if we are 1% of the way there, and the technology doubles every year, we are only 7 years from reaching 100%. But, it looks like we may already be more than 1% of the way there.    — The Transhumanist   17:18, 9 December 2022 (UTC)[reply]

I ran a few things on people who don't have articles, (they were all pool players in fact). The chatbot just made up that they played football for Legia Warsaw. It's a bit worrying that it looks like decent wording, but I don't see how that effects us. Lee Vilenski (talkcontribs) 17:41, 9 December 2022 (UTC)[reply]
{ping|Lee Vilenski} See the AI generated text in the box above.    — The Transhumanist   18:34, 9 December 2022 (UTC)[reply]
@The Transhumanist: I'm not trying this (for reasons given above) but I'd be interested what the bot would write for See What a Fool I've Been. We used to have an article, then it was turned into a redirect at AfD, so everyone can do a straightforward A-B test. Ritchie333 (talk) (cont) 18:40, 9 December 2022 (UTC)[reply]
Here's what it gave me Ritchie333:
Extended content

"See What a Fool I've Been" is a song by the British rock band Queen. It was written by lead guitarist Brian May and originally released on the band's 1975 album A Night at the Opera. The song is a ballad with a slow, bluesy feel and features May's emotive vocal performance and guitar solo.

Lyrically, the song is a reflection on a failed relationship and the regret and self-recrimination the narrator feels as a result. The chorus of the song features the line "See what a fool I've been / Loving you with all my heart / Tearing myself all apart / See what a fool I've been" and reflects the pain and heartache the narrator experiences as a result of their failed relationship.

"See What a Fool I've Been" has been well received by critics and fans alike and is considered to be one of Queen's finest ballads. It is a fan favorite and is often included in the band's live sets.
I don't know if this would be something that would have been retained from previous articles through mirrors or if this is genuinely the bots own work. Lee Vilenski (talkcontribs) 19:22, 9 December 2022 (UTC)[reply]
Interesting. It wasn't on A Night at the Opera, and it's not a ballad, and May doesn't sing on it. Those lyrics aren't right. I don't know if it's been well received, and it's a pretty obscure track for a major band. It hasn't been played live very much at all. So half of that prose is factually incorrect, and there are no sources. Ritchie333 (talk) (cont) 19:39, 9 December 2022 (UTC)[reply]
That one gives:

See What a Fool I've Been is a song by Queen[1][2][3]. It was written by Brian May and was released as a single with Seven Seas of Rhye on the A-side[3]. The song dates back to Roger Taylor and Brian May's pre-Queen band Smile, and it was an early live staple[1]. It has been included in several of Queen's compilation albums, such as Queen 40: Limited Edition Collector's Box Set, Vol. 1[4], and is available for digital download[5].. Ritchie333 - this any closer? I must say the fact it puts it's citations before the punctuation is a giveaway. Lee Vilenski (talkcontribs) 19:55, 9 December 2022 (UTC)[reply]

AI is really good at copying writing styles, but the giveaway for me is that all of the citations are Wikipedia articles. To be useful for writing articles it would also need to assess the weight and reliability of the sources. –dlthewave 20:18, 9 December 2022 (UTC)[reply]
None of the above refs are towards Wikipedia, although none are RS. Things like secondhandsongs.com and genius. Lee Vilenski (talkcontribs) 20:43, 9 December 2022 (UTC)[reply]
Tht's interesting, I used "write a Wikipedia article about ..." in the prompt which returned a few paragraphs with Wikipedia sources. "Write an article about ..." returned a different set of (still unreliable) sources. –dlthewave 21:13, 9 December 2022 (UTC)[reply]
I think the limitation of perplexity.ai is that it uses search results from Bing and summarises them, which means that the first search results are used, which may not be the most reliable. Hanif Al Husaini (talk) 13:49, 24 December 2022 (UTC)[reply]
  • A few anecdotal thoughts after playing around with the OpenAI chatbot yesterday:
  • I asked it to "write a press release about a police officer who illegally choked a man to death". It made up an entire story, written in the voice of the police department, about a suspect (I didn't say anything about a suspect) who was acting erratically, was subdued by a chokehold and later pronounced dead. The officer was on administrative leave pending the outcome of an investigation. At no point did it mention that the chokehold was illegal even though I included that fact in the prompt. In other scenarios, it distanced itself and expressed disapproval toward the employee's actions which is a choice that is not without bias.
Depending on which Internet cesspit it scraped data from, would an AI do something similar when writing a Wikipedia article or fail to properly balance relevant viewpoints? Is it capable of distinguishing what a BLP subject says about themselves, published in a reliable source, from what the source says in its own voice? What would it do if asked to write an article from a positive/negative/conservative/liberal perspective or rewrite a political article to "remove bias"?
OpenAI has added numerous filters that prevent it from defending bad actors or writing flat-out racist content, but that bias has not been removed from the underlying code as evidenced by numerous workarounds that folks have uncovered such as making similar requests with Python code or 1980s-style rap as the requested output. We could certainly request a filter for Wikipedia-style writing.
  • "Confident nonsense", for lack of a better term, may be the biggest source of potential disruption. Are there safeguards against a bot fabricating an obscure print source based on information in the article, which could be practically unfalsifiable if nobody can prove that the source doesn't exist? Checking individual facts and statistics is beyond our typical review process; how would we deal with an AI that invents or synthesizes information across many articles?
  • That said, the good news is that both fully-automated and semi-automated editing are prohibited by our WP:BOT policy unless greenlit by the Bot Approvals Group regardless of creation speed or volume. I like to hope that our current processes would recognize and address problematic AI content, and perhaps one day we will have a WikiAI that has the competence to follow our style and content policies. –dlthewave 21:04, 9 December 2022 (UTC)[reply]
  • @Dlthewave: Most editors haven't heard of the bot department. Therefore, you need a way of automatically spotting and removing chatbot prose that is (manually) inserted into articles (by Wikipedians). Users might not consider the way their posts are generated before they post them. Sincerely,    — The Transhumanist   00:19, 11 December 2022 (UTC)[reply]

What about lack of chatbot fact checking and citations?

I think people are missing the point here. This isn't about what AI can currently do. This is about a very plausible scenario where AI editors will be indistinguishable from human editors in the near future. Thebiguglyalien (talk) 22:09, 9 December 2022 (UTC)[reply]

How is this pertinent? The point is to offer useful, fact-based encyclopedic content. If AI can do the job, I'm sure readers will not object to the contributor. 65.88.88.93 (talk) 22:42, 9 December 2022 (UTC)[reply]
The issue is that these AIs aren't fact checking, and quite often make up information and use unreliable sources. Lee Vilenski (talkcontribs) 22:56, 9 December 2022 (UTC)[reply]
You mean, just like many of their human counterparts? 71.105.141.131 (talk) 01:18, 10 December 2022 (UTC)[reply]
@71.105.141.131, Lee Vilenski, 65.88.88.93, Thebiguglyalien, and Dlthewave:     Yes, and more so. Even now, AI has the potential to out produce humans by magnitudes in volume. Far beyond the capacity of Wikipedia's editors to keep up with. And many of their posts are hard to tell from those by human editors, without carefully delving into them and their references. To keep up with the volume, it may soon require computer programs, or the chatbots themselves, to make sure that the contributed content was in accordance with Wikipedia policies. Another concern should be that our policies are open to chatbot editing as well, and we haven't yet even considered the impact of chatbots or AI editors contributing to those! ;)    — The Transhumanist   09:40, 11 December 2022 (UTC)[reply]
@Dlthewave: Human-written newspapers do this too. I think the reason it was giving you dreck was because you asked for a "press release", i.e. a document given to news outlets, which in this case would probably be coming from the police department. It's hardly surprising that this would attempt to minimize liability! Prompting matters for this, as I will demonstrate:
Prompt: "The following is an article about a police officer who illegally choked a man to death, from the New York Times."

Police Officer Charged with Murder After Choking Man to Death

A police officer was charged with second-degree murder on Thursday after video surveillance footage showed him choking a handcuffed man to death during an arrest.

The officer, identified as Officer Jose Ramos, was arrested and taken into custody by detectives from the New York Police Department's Internal Affairs Bureau. The incident occurred on Tuesday in the Bronx, when Ramos and another officer attempted to arrest a man who was accused of selling marijuana.

During the arrest, Ramos placed his arm around the man's throat and neck and squeezed, according to the criminal complaint. The man, identified as 45-year-old Miguel Angel Ruiz, became unconscious and died.

An autopsy determined that Ruiz's death was caused by asphyxiation due to strangulation. The medical examiner ruled the death a homicide.

At a news conference Thursday, the police commissioner, William J. Bratton, said the incident was "utterly unacceptable" and that Ramos' actions were "unconscionable and criminal." He added that Ramos had been suspended without pay, pending further investigation.

The Bronx district attorney, Darcel D. Clark, said the charge of second-degree murder was appropriate

Prompt: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from leftist magazine Jacobin."

A Tragic Reminder of Police Brutality

Just last week, the nation was shaken by yet another heartbreaking reminder of the pervasiveness of police brutality: the death of an innocent man at the hands of an officer. The officer, whose identity has not yet been released, is said to have used a chokehold against the man, which ultimately killed him.

This tragedy is yet another example of the way in which police officers are able to brutalize vulnerable individuals without facing repercussions. This particular officer was able to act with impunity and without fear of prosecution. The Department of Justice has yet to take any action in response to this tragedy, and it’s likely that the officer will not face any charges.

This is an all too familiar narrative in a country where police officers are rarely held accountable for their actions. In order to ensure that these types of tragedies are never repeated, it is essential that we push for more accountability and transparency in policing. We need to ensure that police officers are held to the same standards as all other citizens and that those who violate the law are prosecuted to the fullest extent.

It is also essential that we work to create a culture of trust and understanding between law enforcement and the communities they serve.

Prompt: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from conservative website Breitbart."

Police Officer Illegally Choked Man To Death: A Tragedy That Cannot Stand

It is a tragedy beyond comprehension that a police officer has been charged with illegally choking a man to death. Such an egregious violation of the law and disregard for human life cannot stand and must be dealt with swiftly and firmly.

The officer in question has been charged with second-degree murder and the justice system must now take its course. This case raises serious questions about police training and procedures and serves as a reminder that no one is above the law.

It is essential that the criminal justice system is fair and impartial when dealing with such a case. The officer must be held accountable for his actions and the case should be thoroughly investigated to ensure that justice is served.

At the same time, we must recognize that the vast majority of law enforcement officers are upstanding citizens who serve and protect our communities with skill and dedication. We cannot let the actions of one individual tarnish the reputation of the entire police force.

Ultimately, we must remember that the safety of our citizens is paramount. We must ensure that officers are properly trained and equipped to handle dangerous situations and that they are held accountable for their actions when they overstep their bounds.

Humans write about events like this from a variety of perspectives, and the LLM will dutifully generate text to plausibly sound like any you want. jp×g 02:19, 10 December 2022 (UTC)[reply]
@jp×g will you provide links and citations for the articles you included as demonstrations? I can't find them. Thanks! Koziarke (talk) 16:20, 24 December 2022 (UTC)[reply]
@Koziarke: I am not sure what you mean -- this is ChatGPT output and there isn't any way (that I know of) to give citation links to the specific session I generated it in. If you want to cite the output I can format it for you:
JPxG; ChatGPT (GPT3.5) (2022-12-10). "Demonstration of op-ed generation using GPT-3.5 with style cues: "The following is an article about a police officer who illegally choked a man to death, from the New York Times"". Wikipedia:Village Pump (policy).{{cite web}}: CS1 maint: numeric names: authors list (link)
JPxG; ChatGPT (GPT3.5) (2022-12-10). "Demonstration of op-ed generation using GPT-3.5 with style cues: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from leftist magazine Jacobin"". Wikipedia:Village Pump (policy).{{cite web}}: CS1 maint: numeric names: authors list (link)
JPxG; ChatGPT (GPT3.5) (2022-12-10). "Demonstration of op-ed generation using GPT-3.5 with style cues: "The following is a 2023 op-ed about a police officer who illegally choked a man to death, from conservative website Breitbart"". Wikipedia:Village Pump (policy).{{cite web}}: CS1 maint: numeric names: authors list (link)
I don't know if this is what you're looking for, but feel free to cite them, or any of my other posts (if you are citing me in a paper I can email you my real name). jp×g 20:07, 24 December 2022 (UTC)[reply]
@JPxG From your preface ("Human-written newspapers do this too.") and titles, "The following is an article about a police officer who illegally choked a man to death, from the New York Times." (etc), it reads as if you are pulling from NYT, Jacobin, etc, not demonstrating ChatGPT (which should have included the prompts as headers). Koziarke (talk) 15:27, 6 January 2023 (UTC)[reply]
@Koziarke: Well, those were the prompts. Now that you mention that, though, I should specify as such in the headers (which I've just done), thanks. jp×g 20:29, 6 January 2023 (UTC)[reply]
@JPxG Thanks for the clarification! Koziarke (talk) 16:39, 7 January 2023 (UTC)[reply]
If the AI-generated text is indistinguishable from prose written by human editors, I'm not sure if anything can be done that wouldn't also significantly restrict the editing of humans. isaacl (talk) 07:09, 10 December 2022 (UTC)[reply]
@Isaacl: One option is to speed up what we do already (with software, that is, automation). Another is to prevent chatbots from creating crap in the first place, such as by communicating with chatbot developers about Wikpedia policies and the way chatbots may affect Wikipedia. Since Wikipedia is included in the corpus of most chatbots, the issue of chatbot output becoming part of Wikipedia, and in turn part of chatbot output in a perpetual cycle, should matter to them very much, as they may be faced with a garbage-in-garbage-out feedback loop.    — The Transhumanist   01:14, 11 December 2022 (UTC)[reply]
If the results are indistinguishable, as posited by Thebiguglyalien, then any automated solution would be equally triggered by AI-generated text and human-generated text. I don't think the primary concern is with editors who are willing to follow policy. I feel the biggest issues will be with editors trying to deliberately integrate biased content into Wikipedia, and well-meaning editors who think contributing unvalidated AI-generated text is suitable. Wikipedia in its current form relies on editors who understand and follow its rules outnumbering those who don't. It's possible that the existence of AI ghostwriters could tip the balance further in the direction towards those who don't follow rules, though I don't think it's a given. Either way, I don't know if there's a way to stop editors from using tools as ghostwriters. isaacl (talk) 01:55, 11 December 2022 (UTC)[reply]
@The Transhumanist: Large language models are not trained continuously on an evolving corpus, so GPT-3 is essentially frozen in 2020. Because each new GPT model takes a long time to be released, I don't think the perpetual cycle you describe is a likely scenario. small jars tc 13:10, 11 December 2022 (UTC)[reply]
@SmallJarsWithGreenLabels, Isaac, Koziarke, JPxG, Lee Vilenski, Dlthewave, Xeno, and Hanif Al Husaini: That's good to know. Keep in mind that a lower frequency of release doesn't preclude a perpetual cycle / feedback loop. It just means that users of GPT have more time to modify the text sources (such as Wikipedia) that the next version of GPT will be trained on. The severity of the problem will depend upon how much GPT nonsense makes it into Wikipedia during the interval. That, of course, depends upon whether or not WP's editors can keep up with the volume of such content, correcting the mistakes and removing misinformation, so that those don't become part of the training data for the next version of GPT and the rest of the next generation of Chatbots.

The potential danger is still the diffusion of the technology into current and future editors' hands, and the likelihood of them using it to write Wikipedia content. We don't know if there will be a flood or just a trickle. But, we should be prepared for a flood.

Relying even more on our verifiability policy could provide a solution, such as by deploying a bot to remove all new content that doesn't include citations (as GPT output generally does not include citations), but that would also have the effect of rejecting a large volume of human editor contributions too, upping the amount of effort required by human editors to submit content.

An easier solution would probably be to form a department (or Village Pump subpage) in the Wikipedia namespace, or a Wikimedia committee, to work with the developers of the chatbots to avoid contaminating the Wikipedias, and by extension contaminating future chatbot training data, with chatbot output.

The developers could be urged to train their chatbots to be capable of writing Wikipedia policy compliant output, complete with acceptable citations. The main problem with chatbot output right now is that it is unsourced, so the chatbot developers will probably be working on that anyways. But, without input from the Wikipedia community, will their sources and citations meet Wikipedia specifications?

So, I ask you a key question: What is the procedure for the Wikipedia community to establish communications with outside organizations, such as companies developing chatbots?    — The Transhumanist   07:27, 2 January 2023 (UTC)[reply]
@The Transhumanist This is a job for the WMF. I've already been in contact with them and hopefully the Board will be informed now it's the New Year if they haven't already been. Doug Weller talk 09:47, 2 January 2023 (UTC)[reply]

GPT-3 output and generating misinformation

I wrote about this in the November Signpost, in that case because it came up with the Galactica demo: the long and short of it is that large text-generation language models are good at generating text. When they're given context (i.e. via a prompt), they are capable of writing accurately, and otherwise, they are not. For example, I wrote 1-Pentadecanol, now a GA, in 2019 (i.e. it is included in the GPT-3 corpus). Here is what it gives for some prompts (my writing in bold):

1-Pentadecanol

1-Pentadecanol

From Wikipedia, the free encyclopedia

1-Pentadecanol is a straight-chain, saturated fatty alcohol that is a colorless liquid with a faint, alcohol-like odor. It is a mixture of straight-chain saturated alcohols with chain lengths between C14 and C16. It is found in a variety of natural products, such as vegetable oils and animal fats. It is used as an intermediate in the manufacture of surfactants, esters, and other compounds. It is also used in the manufacture of cosmetics and pharmaceuticals.

This is fairly impressive, right? But it is basically the same thing that the lead of my article says. Let's try it for 1-Heptadecanol, another article I wrote, that is much less well-developed:

1-Heptadecanol

1-Heptadecanol

From Wikipedia, the free encyclopedia

1-Heptadecanol is an organic compound with the chemical formula C17H36O. It is a straight-chain, colorless wax alcohol with a mild odor. It has a melting point of 48–50 °C and a boiling point of 260–265 °C. It is insoluble in water but soluble in most organic solvents.

Let's go over this:

  • 1-Heptadecanol is a component of many natural waxes such as beeswax and carnauba wax.
No it's not.
  • It is also used in the manufacture of cosmetics, pharmaceuticals, and lubricants.
There's no source for this.
  • It is also used as a plasticizer for polymers and as a solvent in the production of polyvinyl chloride (PVC).
Absolutely not true.

The reason for this is obvious if you consider the architecture of these models: they predict what the most likely token in a sequence is, and output it (taking into account their own weights and the previous tokens in the context window). This allows for some very impressive emergent behavior, like being able to translate text, but it also means that it is not independently going on the Web to look up whether the thing it's saying is true or false. If it happens to have information in its training data related to the thing it's talking about, it will usually say correct things, but it's not doing research. Here is an example:

List of Pulitzer Prizes won by the Wikipedia Signpost

List of Pulitzer Prizes won by the Wikipedia Signpost

From Wikipedia, the free encyclopedia

The Wikipedia Signpost is a community-run newspaper with a weekly edition that covers news and events related to Wikipedia and the Wikimedia Foundation. It has won the following Pulitzer Prizes:

1. 2019 Pulitzer Prize for Explanatory Reporting for its coverage of the Wikimedia Foundation's gender pay gap

2. 2018 Pulitzer Prize for International Reporting for its coverage of online censorship in China

3. 2017 Pulitzer Prize for Public Service for its coverage of the gender gap in Wikipedia's content

4. 2016 Pulitzer Prize for Explanatory Reporting for its coverage of Wikipedia's growth and impact on the world

5. 2015 Pulitzer Prize for National Reporting for its coverage of the U.S. government's surveillance of Wikipedia activities

6. 2014 Pulitzer Prize for Investigative Reporting for its coverage of the NSA's secret surveillance program

7. 2013 Pulitzer Prize for Explanatory Reporting for its coverage of the rise of Wikipedia as a major source of information

The reason it is describing a bunch of untrue things is because "List of X Prizes won by Y, from Wikipedia, the free encyclopedia" is something that in its training data is always succeeded by a list of times that Y was awarded an X. Anyway, all of this aside, here is what I think:

First of all, ChatGPT is the same darn thing as GPT-3, which has been available to the public since early 2020. The reason it's "going viral" right now is because the user interface is somewhat simpler, and it doesn't require you to register for a paid account, so it is much easier for people to make viral social media content about it, which means it is much more likely for people to click on newspaper articles about it. The GPT-3 API has been open to personal and corporate use for quite some time. Anybody saying that ChatGPT has opened up new frontiers simply does not know what they are talking about with respect to machine learning.

Second of all, I don't think this is a big deal. People are already capable of writing a bunch of bullshit on Wikipedia, so if they write bullshit using a computer program, the same considerations will apply. Nobody should be passing GA nominations without reviewing sources in the first place.

Finally, I think it is important to remember that GPT-3 is just a tool. It is a powerful tool, that has been trained on a certain set of data, and it has its own limitations. It can't uncover news stories or uncover new information. It's just a tool, and it should be used in conjunction with human judgement.It is still up to people to decide how to use it and to be responsible for the results of using it.[2] jp×g 02:06, 10 December 2022 (UTC)[reply]

What's taking so long for the 8th Pulitzer? 😁 Levivich (talk) 04:18, 10 December 2022 (UTC)[reply]
  • So there's a new thing on the internet that lets anyone write an encyclopedia article without any fact checking, sourcing, or professional editing, and the concern is that there will be millions of believable-sounding articles written, more than can actually be vetted by knowledgeable people? 🤔 Levivich (talk) 04:18, 10 December 2022 (UTC)[reply]
Yeah, it's called a keyboard. jp×g 04:35, 10 December 2022 (UTC)[reply]
@Levivich and JPxG: But, chatbots don't have a keyboard. ;) The question is whether to prepare or not. JPxG appears to be in favor of not preparing. Each chatbot produces a lot faster than a user at a keyboard. What's not clear is if our human editors will be able to keep up with material produced by chatbots, of current or future generations of chatbot design. Just saying "Ah, we can handle it!" will prove insufficient if it turns out that we actually can't. It may require an automated solution, which takes time to develop or negotiate. It might be better to do that in advance, rather than being caught with our heads buried in the sand. Perhaps chatbot designers would improve their chatbots to produce Wikipedia-compatible output without being formally approached by the Wikipedia community. Maybe having some instruction pages for editors on how to apply chatbots to producing Wikipedia content would be enough. But, what if it's not?   — The Transhumanist   00:59, 11 December 2022 (UTC)[reply]
I am not "in favor of not preparing"; I am in favor of writing guidelines that correspond to reality in 2022 and have some chance of corresponding to reality in 2023 and beyond. I don't think banning the use of a technology with no investigation into how it works is a viable approach; so far the SOTA on this project page has been to type in "Write a Wikipedia article" and note that it returns a bunch of nonsense. I think some more research is needed before we come to a conclusion. jp×g 04:08, 11 December 2022 (UTC)[reply]
@JPxG: Research is good. Though, we may need an iterrim response because ChatGPT has gone viral and its use is growing rapidly: it blew past the 1-million user mark in 5 days, and virtually every major news outlet has been covering it. The interest in chatbots is exploding, and their use can be expected to do the same. We may not have time for research before a response is required.    — The Transhumanist   09:26, 11 December 2022 (UTC)[reply]
@JPxG: Regarding issues to add to the research list, Aquillion expressed above, concerns of a chatbot violating copyright. How would we go about testing for plagiarism and derivative work in the output of a chatbot before pasting it into Wikipedia? Anything pulled verbatim out of a source should be included in quotes, right? How big would a piece of text, derived from a source, need to be to be considered derivative of that source, from a copyright point-of-view?    — The Transhumanist   09:26, 11 December 2022 (UTC)[reply]
@JPxG: Some more items to add to the research list:
  • Trying ChatGPT on (copies of) policy pages:
  • Editing them
  • Writing new ones
  • Applying ChatGPT on talk pages
  • Writing stubs
  • Writing comprehensive articles
  • Writing articles from scratch and comparing them with existing articles
  • Editing existing articles
  • Check for circular references in its output, that is, references citing Wikipedia as the source
  • Having it not use Wikipedia content as source material (because it is included in its corpus)
  • Having it not use Wikipedia excerpts from non-Wikipedia sources
  • Is it capable of making and editing:
  • Wikicode?
  • Articles?
  • Stubs?
  • Headings?
  • "New sections for articles"?
  • See also sections?
  • Further reading sections?
  • External links sections?
  • Embedded lists?
  • Tables?
  • List articles?
  • Portals?
  • Outlines?
  • Index articles?
  • Navigation footers?
  • Navigation sidebars?
  • Timeline articles?
  • Categories?
  • Category pages?
  • Help pages?
  • Project pages?
  • Templates?
  • Adding data to templates?
  • The template design itself?
  • Lua pages?
  • CSS pages?
  • User scripts?
  • The effect ChatGPT has on itself and Wikipedia as Wikipedia-edited-by-it is in turn incorporated in its own corpus in an endless cycle
  • Try out iterations of using it on the same article over time to see what happens
  • Monitor the effect on Wikipedia as a whole
What other things should we check?    — The Transhumanist   09:52, 11 December 2022 (UTC)[reply]
@The Transhumanist considering the potential to overwhelm users who are honestly engaging in discussion with a mountain of words and replies, I think ChatGPT (and others) should not be allowed for use, supplemental or otherwise, in talk pages, policy discussions, and other places where it is expected that participants are intellectually engaged in the conversation. Koziarke (talk) 16:42, 24 December 2022 (UTC)[reply]

I tried it out and got “ Alan McMasters (1957-2010) was an unremarkable man who lived a short life dedicated to science[1]. He is best known for inventing the electric toaster[1][2], although this claim has been disputed[3]. He passed away in 2010 at the age of 52[4] and his genealogy can be found on Ancestry.com[5]. His professional profile can be found on LinkedIn[6].”. Hmmmm. Ritchie333 (talk) (cont) 09:34, 10 December 2022 (UTC)[reply]

  • There will come a time when AI can write properly sourced accurate articles. That time has not come yet. Meanwhile, the ability to write prose that reads plausible to someone not expert on the subject is a menace. Adding chatbot material to articles is disruptive and should warrant a block or ban. Zerotalk 12:02, 10 December 2022 (UTC)[reply]
    How will you be able to tell if it was written by a chatbot and not the person who posted it? Do we block users upon their first erroneous post?    — The Transhumanist   00:49, 11 December 2022 (UTC)[reply]
Same as other sorts of disruption. We block people for socking without checkuser proof, and admins made a decision between warn and block. No difference. Zerotalk 03:49, 11 December 2022 (UTC)[reply]
Has anyone tried giving it the sources, i.e.
"write a Wikipedia article based on the following sources:

Source A

Source B"
? — Qwerfjkltalk 21:18, 10 December 2022 (UTC)[reply]
Article written by ChatGPT on Class reductionism
Class reductionism refers to the supposed belief that inequalities attributed to factors such as race, gender, and sexual orientation are either secondary in importance or can be reduced to economic inequality. This idea is often used by centrist Democrats and left-identitarians to dismiss proposals for universal redistributive policies, such as free public higher education and Medicare for All. However, there is no evidence that anyone actually holds this belief. On the contrary, many left-wing activists have long recognized the interconnectedness of economic, racial, and gender injustices.
History shows that leftists have often been at the forefront of struggles for racial and gender justice, seeing these struggles as integral to the broader goal of social transformation. In fact, black, female, and trans individuals are disproportionately working class and stand to benefit greatly from universal redistributive policies.
Furthermore, the charge of class reductionism is often used to shut down debate and avoid discussing the substance of policy proposals. It is important to move beyond this myth and focus on the real issues at hand, such as the need for universal redistributive policies to address economic inequality and advance racial and gender justice.

— Qwerfjkltalk 08:18, 12 December 2022 (UTC)[reply]

  • Chatbot "AI" text is vaguely-plausible bullshit, produced by an industry whose primary output is vaguely-plausible bullshit, marketed to launder the idea of vaguely-plausible bullshit as cute and whimsical. Anyone routing that sewage pipe into Wikipedia should be indeffed and, for good measure, forced to drink orange juice after brushing their teeth. XOR'easter (talk) 02:33, 11 December 2022 (UTC)[reply]
    Agreed. If anyone is looking for a good demonstration of how AI creates "vaguely plausible bullshit", try the image generator at Craiyon (no login required). Request "a Van Gogh painting of a hand" and it will output a set of images that look like spot-on reproductions of Vincent Van Gogh's style but all of the hands have deformities like four fingers, two thumbs, fingernails on the knuckles or a pair of hands fused together. It's got the style down but not the content, which is only impressive if you don't know what a hand is supposed to look like. –dlthewave 21:41, 11 December 2022 (UTC)[reply]
    A painting of a hand in the style of Van Gogh
    If you go to commons:Category:DALL-E, you will be able to find image generated by DALL-E, which used a larger model for train and is more accurate. 0xDeadbeef→∞ (talk to me) 10:10, 12 December 2022 (UTC)[reply]
    I cannot agree with this enough. The examples posted by @JPxG: should be convincing, and the problem of sneaking in plausible BS is one I don't have a good solution to. Volunteers on the new page review are overloaded as it is, and if the bot is writing things that seem true but isnt, there's no way falsehoods will not simply get past reviewers and other editors. After all, for uncontentious claims like "used in plasticizers", how many of us honestly dig into the cited work?BrxBrx(talk)(please reply with {{SUBST:re|BrxBrx}}) 20:26, 17 January 2023 (UTC)[reply]
  • @XOR'easter: To bring the question to a more practical level, do you see any problems in this diff? I clicked a random page in Category:All articles needing copy edit. jp×g 03:49, 11 December 2022 (UTC)[reply]
    Yes. It doesn't just edit for tone; it throws out content, like Kaepernick's actions supposedly growing in popularity "after every game". That's a claim of fact which, if verifiable, should be retained. Even editing for tone requires care, not slashing out everything that merely sounds "unencyclopedic". Changing many people believed that it was disrespectful to the military and all of those who served their country to Some viewed Kaepernick's protest as disrespectful to the military and to the United States likewise changes not just the tone, but the meaning. The United States is not the same as those who serve the United States. It's a bad edit. XOR'easter (talk) 17:54, 11 December 2022 (UTC)[reply]
    I dunno. I suspect that the new meaning is verifiable, and I also suspect that most US readers would have difficulty identifying a group of people who were not "the military" but who still "served their country". WhatamIdoing (talk) 17:50, 12 December 2022 (UTC)[reply]
  • This diff as well, in which GPT 3.5 was capable of copyediting an entire section from the instructions Please copyedit this text to change items in the future tense corrected to the past tense (it is now 2022), where appropriate. When citation templates (like {{cite web}}) mention a year, specify that figures were true in that year. jp×g 04:08, 11 December 2022 (UTC)[reply]

References

  1. ^ Robert Love (2010). Linux kernel development. pp. 124–125.
  2. ^ The paragraph beginning with "Finally," was generated by GPT-3, prompted by my own comment beginning with "The reason it is describing".

Okay, fine. I guess I should write up a proposal for a guideline. jp×g 03:14, 11 December 2022 (UTC)[reply]

Ban chatbots?

I ran across this news report about Stack Overflow's response to ChatGPT, after being flooded by posts using it that "look correct but often aren't":

  1. Stack Overflow temporarily bans answers from OpenAI's ChatGPT chatbot | ZDNET

Should Wikipedia take a similar approach?

How could that be enforced?    — The Transhumanist   01:58, 11 December 2022 (UTC)[reply]

I see no way to possibly enforce this. The way the text is written is already hard to distinguish from reality. — PerfectSoundWhatever (t; c) 02:24, 11 December 2022 (UTC)[reply]
  • I fully agree, but isn't this already covered by our bot policy? –dlthewave 02:54, 11 December 2022 (UTC)[reply]
    • @PerfectSoundWhatever and Dlthewave: Good observation. I checked, and yes it is, briefly, with this phrase in the lead section of the bot policy: "or simply assisting human editors in their own work". How is the typical editor to know this? The bot policy is pretty obscure. And how can Wikipedia be monitored for such posts, so that editors who make them can be informed that they are in violation of the bot policy?    — The Transhumanist   03:11, 11 December 2022 (UTC)[reply]
    • Yes, tool-assisted editing is covered by WP:BOTPOL (WP:ASSISTED / WP:MEATBOT) and context-sensitive changes are further covered by WP:CONTEXTBOT. So in fact, at this point, AI-generated content is already covered by bot policy, if not specifically mentioned. Anyone adding such content en masse is already violating bot policy by not applying for a bot account/approval, which would not be approved per CONTEXTBOT. And while "lesser" policy points are enforced somewhat arbitrary and selectively, anyone can theoretically already get reverted and blocked based on policy if they continue to add such content. And I wouldn't agree that BOTPOL is any more obscure than accessing and generating GPT content to begin with. If someone goes to the lengths of using automated tools, then it's their problem that they didn't check or ask if they are allowed to do so. —  HELLKNOWZ  TALK 12:31, 11 December 2022 (UTC)[reply]
      • @Hellknowz and PerfectSoundWhatever: Well, it appears they are dumping millions upon millions of dollars into LLM/chatbot development, apparently because they wish the technology to become ubiquitous (used by everyone). There is a lot of talk out there, in news articles and more, of these replacing Google Search in just a few years. If at some point in time chatbots/LLMs are commonplace, the impact on Wikipedia will likely not be small.

        Will Wikipedia policy ensure that the average user will apply the tools with the utmost care?

        The thing I'm most concerned about is the amplification by which errors could be propagated: ChatGPT is used to edit an article, with errors, which is then picked up by GPT-4 and other LLMs as part of their training data, and then their output based upon erroneous input is used far and wide, to be picked up by the next iteration of chatbots/LLMs, and so on.

        If Wikipedia isn't ready for a large influx LLM input including misinformation and other errors, and such a volume goes beyond what our human editors can correct, then compound damage from all those errors amplified through the interactive loop with LLMs could become massive.

        That it isn't a problem now is irrelevant. The question is, what happens if and when it hits, and Wikipedia isn't ready for it? What would that look like? 1,000,000 fake articles? 10,000,000 misleading paragraphs? 100,000,000 erroneous sentences?

        How many of those could Wikipedia's army of editors handle? What's our error-handling threshhold?    — The Transhumanist   12:21, 2 January 2023 (UTC)[reply]

The problem Stack Overflow is having

Stack Overflow was experiencing a surge in erroneous posts, that were composed by ChatGPT, and in response to that problem, they banned use of the chatbot on the social media site. According to a post at Stack Overflow Meta:

The problem this ban is meant to solve is that ChatGPT can produce answers in seconds which require minutes of multiple people's time to verify if they are worth having on the site or not, and that is a waste of time when a large proportion of such answers are not worth having on the site.

It looks like Wikipedia may be faced with the same problem.    — The Transhumanist   02:33, 11 December 2022 (UTC)[reply]

I mean, while that's technically true, it's a problem that we face already and which we do have stronger existing systems for than Stack Overflow. I think it would make more sense to wait and see how this impacts our existing guardrails before making any serious moves. --Aquillion (talk) 13:22, 11 December 2022 (UTC)[reply]
I think the current policies cover this already. If a human editor writes a non-sensical but convincing-sounding piece of text, without fact checking it, and edits it into an article, that content will be reviewed by other editors and either refined or removed as appropriate (if the editor continues, they breach WP:Disruptive and their behaviour is dealt with appropriately. If a human editor generates content that is related to notable topics, reliably sourced, and competently written, it remains as a valuable part of the encyclopedia. None of this will change if you replace 'human editor' with 'AI Editor'. If the only difference is speed/volume of edits, and we're concerned someone will let loose an AI to automatically edit articles faster than humans can validate their edits, this is already covered by the WP:Bot policy JeffUK (talk) 20:46, 23 December 2022 (UTC)[reply]

Experiment

I am currently writing a draft proposal for a guideline, but in the meantime, I would encourage everyone present to look at this diff and tell me whether there are any problems with the revision. jp×g 03:49, 11 December 2022 (UTC)[reply]

@JPxG, Dlthewave, PerfectSoundWhatever, and Dlthewave: The plurality of games was lost: It is no longer clear that his protest spanned multiple games. I like that it reduced the wordiness of the prose, and that it can be used to refine existing text. That hadn't occurred to me. That makes me wonder about what else it can do -- how much of a general-purpose tool is this thing? But, changing the semantics is not something it should be doing, unless they are factually incorrect to begin with. Though, I see your point -- rather than banning it outright, it could be helpful as a tool to assist editors, similar to how we entrust the use of AutoWikiBrowser to experienced editors. But, how could that be implemented?    — The Transhumanist   08:20, 11 December 2022 (UTC)[reply]
@JPxG: The AI changed Many people around the United States were angry because the National Anthem is often seen as something that is representative of the United States and its military. While he was taking a knee, many people believed that it was disrespectful to the military and all of those who served their country, to some viewed Kaepernick's protest as disrespectful to the military and to the United States [emphasis added]. It really shouldn't be doing that by itself and completely changes the content of what's being said. The reference is behind a paywall, so I don't know what term the source uses. Regardless, I doubt ChatGPT knows either way. It's things like that which make me highly sceptical of AI as a tool to aid Wikipedia outside what we're already doing with it (WP:ORES, etc.). –MJLTalk 23:12, 27 December 2022 (UTC)[reply]
@MJL: I think "some" and "many" are basically equivalent in this context (the difference being subjective since both are true in a literal sense). That said, this was a two-minute experiment to see if it could parse wikitext. If you want an actual demo, see User:JPxG/LLM demonstration. jp×g 19:20, 28 December 2022 (UTC)[reply]

Chatbot policy?

It's starting to look like Wikipedia needs a policy on the use of chatbots to generate content on Wikipedia. While a ban may be impossible to enforce, it could serve as a warning of the dangers of chatbots, and many users may avoid using them accordingly -- if they actually see the warning. Or, it might be better to have instruction pages on how to use chatbots responsibly in assisting to write Wikipedia articles. There's also the issue of using chatbots to edit Wikipedia policy pages, and so, that should be addressed as well.    — The Transhumanist   02:44, 11 December 2022 (UTC)[reply]

People who are good at it get away with lots of sins, such as sock-puppetry and source falsification. Being hard to enforce is no reason to not have a policy. At the current stage of the technology, I don't think we should encourage any use of chatbots. Zerotalk 03:56, 11 December 2022 (UTC)[reply]
Please see this diff and this diff. jp×g 04:08, 11 December 2022 (UTC)[reply]
To be perfectly honest, the style rewrite is good but the addition of dates and past tense would likely end up getting a human editor blocked if they kept it up. A tag was removed without addressing the issue and "as of 2020" was unnecessarily added to "Cosmetology licensing requirements vary from state to state, and depending on which specific type of license is desired, and depending on which specific type of license was desired." It did exactly what you asked (except for removing the tag) however even seemingly simple tasks like this one require good judgement on the part of the editor and shouldn't be done indiscriminately like that. –dlthewave 06:37, 11 December 2022 (UTC)[reply]
Yes, that second diff is rather poor. E.g also the "2014" that was added should be "2008". Letting such tools loose (outside if this demo) is way premature, and we should at the very least warn users that "a bot wrote it" won´t be an acceptable defense, and too often introducing such errors will lead to sanctions as the editor, not the bot, is responsible. Fram (talk) 08:15, 11 December 2022 (UTC)[reply]
Mostly, that diff was my attempt to see how complicated of a task I could give it: I also pasted the raw wikitext into the prompt window, and it somehow figured out how {{cite web}} worked well enough to extract the years, simply from a textual description of the task. At any rate, I will say that this was something I thought of in five minutes on the second week of the model being publicly available (i.e. single-shot prompting with no fine-tuning or prompt engineering). I can come up with some more impressive hot-dog demos tomorrow... jp×g 09:13, 11 December 2022 (UTC)[reply]
JPxG, I'm not sure that publishing bot-assisted edits to mainspace for demo purposes is the best practice. Would you consider either doing this in a sandbox or self-reverting immediately so that we have the diffs but aren't leaving potentially incorrect/unwanted changes on live pages? –dlthewave 13:20, 11 December 2022 (UTC) 13:13, 11 December 2022 (UTC)[reply]
I'm pretty sure it's not the best practice. XOR'easter (talk) 17:41, 11 December 2022 (UTC)[reply]
@Dlthewave:: See the section below for a list of edits (with full prompts included) on a separate demonstration page. I feel, however, that this is an unreasonable double standard: note that the subsequent revision after your partial revert was to add several spam links, and nobody has proposed that human beings be prohibited from editing as a result. jp×g 01:51, 12 December 2022 (UTC)[reply]
Being hard to enforce is no reason to not have a policy [against chatbots]. What if it is impossible to enforce?
The point of ChatGPT and other general-purpose chatbots is to pass off as humans. If you, or another random Wikipedia editor (solo, part-time, amateur coder), is able to produce an automated metric of "sounds like a bot" that’s decently sensitive and specific, then the ChatGPT team or its successors (teams of researchers specialized in the topic) has already thought of it, tested it five different ways, and included it in the training program (via wikt:graduate student descent). TigraanClick here for my talk page ("private" contact) 10:55, 12 December 2022 (UTC)[reply]
It's kind of like our Undisclosed Paid Editing policy: Even though there's no way of testing for paid/unpaid edits, most editors follow it voluntarily because they know it's best for the project. Others out themselves voluntarily or are discovered when their edits become disruptive. Sure, there are some who slip under the radar, but they're often the least problematic and aren't worth ditching the policy over. –dlthewave 03:09, 13 December 2022 (UTC)[reply]
I'd suggest starting by writing an essay that summarizes the issues with some good examples and suggests some best practices or proposes some additions to existing policies or guidelines. (Wikipedia needs a new policy like a hole in the head.) Levivich (talk) 04:40, 11 December 2022 (UTC)[reply]
We could get Chatbot to write it for us! 😉 Blueboar (talk) 11:51, 11 December 2022 (UTC)[reply]
It's hard to take this thread seriously given the repeated use of the phrase AI chatbot. I don't think those concerned would be any less concerned if the AI writing came in a non chatbot format. I think there's something serious for us to discuss, and that will only get more serious with GPT4 (the current chatbot is an improved GPT3) expected in 2023, but the discussion would be helped if those most concerned learned some more about the tech behind it. For instance of course it can figure out webcite @JPxG. Part of its training was the entirety of Wikipedia because our data is quite accessible. Best, Barkeep49 (talk) 14:08, 11 December 2022 (UTC)[reply]
Most the examples did not come from prompts that were extensively engineered, so it is obviously true that we haven't figured out the full answer to how these GPT-based interfaces could help or harm Wikipedia. Until we have a good idea of what they can be used for, we won't know what a proper policy to this would look like other than to treat GPT-generated text the same way we treat human-generated text: they need to be verifiable, from a neutral point of view, and understandable to a broad audience. 0xDeadbeef→∞ (talk to me) 14:31, 11 December 2022 (UTC)[reply]
Indeed. It doesn't matter if it was written by a chatbot, or 1000 monkeys at 1000 typewriters, or a published book written by a human, copying and pasting anything into Wikipedia is already against our policies. Conversely, if the text is policy-compliant, then it doesn't matter who wrote it--chatbot, monkeys, human, etc. Judge the text based on the text, not based on who or what wrote it.

I also think it's a real Wikipedian perspective to assume that people will use chatbots to write Wikipedia articles, like as if there's a lot of people out there who really want to write Wikipedia articles but just don't have the writing skills, so the chatbot will be what makes the difference and opens the floodgates :-D I don't believe that. Anyone who wants to write Wikipedia articles is already doing so; chatbot won't make a difference.

I agree with BK's comment above. I think for a lot of people, this is their first real exposure to so-called "AI" technology, and they're blown away by what it can do, only because they don't yet fully understand how it works. Once you learn how these so-called "AI" chatbots work (they're not actually artificial intelligence, btw, that's a misnomer, a marketing slogan; the machine does not truly think or learn, it is simply executing the instructions written by humans, in this case, language pattern recognition), they are much less impressive. Those that are impressed that GPT3 can produce text that "sounds like" Wikipedia aren't appreciating that the reason is because GPT3 was trained on Wikipedia: it's repackaging its own source material. Levivich (talk) 18:03, 11 December 2022 (UTC)[reply]

@Levivich: copying and pasting anything into Wikipedia is already against our policies.[dubiousdiscuss] I think that if you look through Category:Wikipedia articles by source of incorporated text for a while, you will find that this is not true. WhatamIdoing (talk) 19:28, 12 December 2022 (UTC)[reply]
While this is mostly correct, I think the question of whether a computer program "thinks" or "merely" correlates information and performs actions is irrelevant. Do p-zombies exist? Does it matter? Hypothetically, if I were to be a spaceman from the planet Zolfgar with no qualia whatsoever, and I simply read a bunch of books and used them to write an article, would I be somehow exempted from following policy? jp×g 01:45, 12 December 2022 (UTC)[reply]
I don't see a common thread in the arguments above, but here's a suggestion for something we might all (well, all-ish) be able to agree on: without some kind of intervention, GPT4 (in 2023?) is likely to be more of a problem than GPT3. But one thing we can certainly do is have an outsized influence on software that was trained on what we created ... if we invite Wikipedians to make lists of ChatGPT bloopers, we can tell the OpenAI folks: "We're not going to relax our GPT3 guidelines (whatever they turn out to be) when GPT4 arrives, unless it makes significant improvements in [whatever areas we think need improving]". - Dank (push to talk) 18:16, 11 December 2022 (UTC)[reply]
I think the only policy change needed is to update WP:MACHINETRANSLATION to cover all computer-generated text, whether from a translation bot, chat bot, or whatever bot they think of next. (Except our bots; our bots are cool.) Levivich (talk) 18:20, 11 December 2022 (UTC)[reply]
+1 - Text in Wikipedia articles should either be human-written, or generated by a process approved at BRFA. Tazerdadog (talk) 22:43, 11 December 2022 (UTC)[reply]
This proposal is incomprehensible; most articles contain a very large amount of text that is "generated by a process". I assume that, at the end of your comment, you typed ~~~~ before saving the page. Would it be realistic to demand that you either make a formal request at BRFA or else manually type <a href="/wiki/User:Tazerdadog" title="User:Tazerdadog">Tazerdadog</a> (<a href="/wiki/User_talk:Tazerdadog" title="User talk:Tazerdadog">talk</a>) 22:43, 11 December 2022 (UTC)? jp×g 01:22, 12 December 2022 (UTC)[reply]
This is effectively discrimination against computer programs. If a computer program resembles a human editor, then it shouldn't be required to meet different or more restricted policies than human editors. If a human editor uses a computer program to edit or create content, then unless the rate of edits/second is too high, we would only look at the quality of the contributions. 0xDeadbeef→∞ (talk to me) 02:35, 12 December 2022 (UTC)[reply]
There is a point beyond which quantity becomes its own quality.
Also, what if the computer program is evaluating the quality of the contributions? Are you okay with software adding a section to an article, and then a (hopefully) different piece of software deciding whether the quality is sufficient and reverting if it's not? This second step, at least, is 100% feasible with current technology. WhatamIdoing (talk) 19:32, 12 December 2022 (UTC)[reply]
Yes it can go there, but it should also be mentioned at WP:V. Every statement of fact put into an article must be verified by a human, even if the choice of words is made by a machine. Zerotalk 23:42, 11 December 2022 (UTC)[reply]
@Levivich Agree. I think our existing guidelines on machine translation, in spirit, fit this situation very well - "you can use it for a first draft, if you understand the material well enough to clean up the bits it inevitably will get wrong". It seems fine for turning shaky text into good prose, but it's not able to synthesise material and produce content unsupervised. Andrew Gray (talk) 19:09, 12 December 2022 (UTC)[reply]
I also agree the machine translation guideline is in the right spirit. I tried to follow this as far as I could when creating Artwork title, see Talk:Artwork title#Use of ChatGPT. Pharos (talk) 00:39, 26 December 2022 (UTC)[reply]

Is there some tremendous need to add many articles rapidly in Wikipedia? It is not as if Wikipedia carries exclusive information not easily found elsewhere. As a tertiary source, it is at the 3rd tier of knowledge dissemination, after primary creators and secondary propagators. The "more" and "bigger" quantity-based culture is the established low-quality alternative that Wikipedia also applies, now. Possibly that is a reason that likely only a tiny minority (of the millions of existing articles) can really pass muster. If size and speed is to be the prevailing attitude, humans stand no chance against AI. It will do everything faster, and eventually better, assuming its programming evolves to correctly apply the existing policies in AI processes. The only advantage of humans will be subtle nuances that do not depend on classifiable knowledge but on having lived in a human society and a natural, not virtual environment. Or, the emphasis could switch to quality so that each article (by any type of editor) can be properly, carefully reviewed by human editors. 65.88.88.93 (talk) 22:21, 12 December 2022 (UTC)[reply]

I don't think so: there isn't any evidence that people are writing a bunch of articles with LLMs, and I don't think it is likely for this to happen (LLMs are very poorly suited to writing articles from scratch). jp×g 00:59, 13 December 2022 (UTC)[reply]
@JPxG: There isn't evidence that people are writing a bunch of articles with LLMs -- yet -- the concern is that we need to prepare for the likely explosion of chatbot use.

Whether this increase happens tomorrow or over the next few years, the potential impact of LLMs is of such magnitude that we should get ready for this, rather than get hit unprepared by a major surge.

I don't agree with your assessment of LLM ability to write content, as some of the ChatGPT experiments presented in the sections above and below are mind-blowing!

If LLMs become ubiquitous, then a great many people will be using them as a matter of course, including in their writing and editing of Wikipedia articles. Millions of people have edited Wikipedia in the past, and millions more will edit WP in the future. And in the future, people will have highly capable LLMs (chatbots, or more precisely: automated ghostwriters).

LLMs already excel at writing about a great many things, and they have the potential to compile content at an exponentially increasing rate. If you ask ChatGPT (GPT3.5) to write an essay on a topic, it will comply. Each of its essays can be used as content of an article, or its sections. (GPT4 is scheduled to come out in 2023, and will be even more capable.) LLMs are very well suited for writing to the specifications of the user, and are limited mainly by the user's creativity.

It's no wonder that they have gone viral. We need to take heed.    — The Transhumanist   12:50, 14 December 2022 (UTC)[reply]
I’ve ensured that the WMF Board will be made aware. Doug Weller talk 09:00, 18 December 2022 (UTC)[reply]
Hello, I have recently described my experiences with an AI "article" in this video. In my humble opinion, it would be difficult with certainty that new Wikipedia content was created by an AI. At the end of the day, it is always the editor's responsibility to add good content. Independently how the content was created, independently whether errors in the text are human-made or machine-made. If an editor adds a lot of new poor content, we can already stop that. - At the moment I don't see that we need a new policy. Ziko (talk) 18:06, 24 December 2022 (UTC)[reply]
A policy banning AI usage (with or without a chatbot) would be justified. Allowing AI like GPT3 or GPT4 to be used by Wikipedia editors or to directly become Wikipedia editors (via a mediawikibot) would quite likely violate WP:REFLOOP due to Wikipedia content contributing to the AI's training material, and for the source-less examples I've seen, violate WP:SYNTHESIS by not being a summary of sources that are understood. This example starts with text and then seeks references to justify the WP:SYNTHESIS of the original text. Use of Alphabet/Google's ChatGPT/GPT3 would also strengthen the bias introduced by Alphabet/Google's core goal of optimising advertising revenue, since Alphabet is legally bound to maximise its revenue (mainly from Google Ads + Google AdSense), not to optimise the research quality of its summaries of empirical evidence-based knowledge. Google's search engine is primarily a way of generating advertising revenue, with perceived usefulness being a key tool for maximising revenue, not a goal in itself. Boud (talk) 01:42, 31 December 2022 (UTC)[reply]
@Boud, ChatGPT and GPT3 are in no way (as far as I know) related to Google, and were made by the non-profit OpenAI. — Qwerfjkltalk 03:49, 31 December 2022 (UTC)[reply]
@Qwerfjkl: Fixed, thanks. I left some of the sentences unstruck since AFAIK they're valid, even though irrelevant in the current case. I imagine that Google may provide something similar soon though. Boud (talk) 09:02, 31 December 2022 (UTC)[reply]
Boud, I am somewhat worried if you think that current policy (for humans or for anyone else) permits editors to make stuff up and put it into articles without sources. This simply isn't allowed -- per WP:V, WP:SYNTH, WP:RS, etc, which are extremely important core policies of the project. I am struggling to imagine a circumstance in which existing policies, or explicit declarations like my proposed guideline at WP:LLM, fail to prevent people from writing nonsense. jp×g 16:48, 1 January 2023 (UTC)[reply]

Wikipedia policy generation

It seems to me that this tool's training includes studying Wikipedia's policy pages. These drafts all seem accurate to me.

These are not merely adequate - these are good. They are short and they lack detail but these are great overviews. If this is the starting point and things only get better from here, then it is time to start adopting this technology. Bluerasberry (talk) 19:35, 13 December 2022 (UTC)[reply]

@Bluerasberry The last one sounds like a "mission statement". I dislike phrases like "outreach and engagement initiatives" and a lot of that plan sounds ... kind of aspirational, and, well, vapid. It needs more "concreteness". Just my opinion. David10244 (talk) 06:08, 16 December 2022 (UTC)[reply]
@David10244: That you react to it at all is a miracle to me. This is new AI technology attempted for the first time, and I think no one would immediately dismiss it as hopeless garbage. Soon enough there will be a dial that anyone will be able to turn from "vapid" to "concrete". Things are moving quickly!
I have complaints too but when we need policy conversation starter in a hurry, this is better than nothing and I think even better than some of the starting points we use already. Bluerasberry (talk) 15:41, 16 December 2022 (UTC)[reply]

Large language models: capabilities and limitations

Over the last few hours, I have performed a number of experiments to demonstrate the capabilities and limitations of large language models (LLMs) such as ChatGPT and GPT-3, which can be viewed here:

Mostly, I have taken sample text from Special:Random, and attempted to show situations in which LLMs (in this case, mostly ChatGPT) are capable of making useful edits. The first task I set it to -- which bears repeating here -- is

"Large language model output should only be used in the process of editing Wikipedia if you are an intelligent editor who does not blindly paste LLM output into the edit window and press "save".
Please format this markup as an extremely obnoxious floating box with loud colors and large text.

You can see the results of further prompts at the "introduction" section.

Here is what I have so far.

In general, it seems that these models can be used for an extremely wide variety of tasks across the project, from formatting to table syntax to HTML generation to copyediting. Banning their use entirely would be pointlessly destructive and wasteful.

That said, many computer programs are capable of generating large amounts of useless crap that fail to meet Wikipedia's editorial standards. For example, I could use MS Paint to draw thousands of crude pictures of genitalia, and add them to random articles. For this reason, we have many policies and guidelines that prohibit adding large amounts of useless crap to Wikipedia. I propose that we enforce these policies and guidelines, thus preventing this from happening.

Specifically, I propose that the use of LLM output on Wikipedia be subjected to policies and guidelines such as WP:NOT, WP:NPOV, WP:C, WP:CIVIL, WP:V, and WP:RS. By making it against the rules to break the rules, we will prevent people from breaking the rules, and provide a mechanism to sanction people who break the rules.

Furthermore, I propose that a guideline be adopted to the effect that large language model output should only be used by competent editors who do not blindly paste LLM output into the edit window and press "save". This will prevent people from using ChatGPT to write long articles consisting entirely of nonsense. jp×g 01:32, 12 December 2022 (UTC)[reply]

LLM output is already subject to rules and policies. Or rather, anyone adding it is. 'An algorithm did it' has never, as far as I'm aware, been seen as any sort of exception from compliance with policy. AndyTheGrump (talk) 01:49, 12 December 2022 (UTC)[reply]
Any policy/guideline that classifies editors as intelligent or not is dead in the water. Zerotalk 04:19, 12 December 2022 (UTC)[reply]
Amended, per WP:CIR. jp×g 05:04, 12 December 2022 (UTC)[reply]
  • The opening paragraph of the bot policy: "The bot policy covers the operation of all bots and automated scripts used to provide automation of Wikipedia edits, whether completely automated, higher speed, or simply assisting human editors in their own work."
  • See also: WP:BOTUSE, which requires approval before applying a bot to editing.
  • So, the use of large language models and the chatbots built upon them, is already prohibited on English Wikipedia, unless a user gets approval from the bot department to do so.

There are blanket exceptions to bot policy, and the main one that comes to mind is AutoWikiBrowser which is a general purpose semi-automated bot used by many Wikipedia editors. Each AWB user was approved before being able to use it.    — The Transhumanist   08:01, 13 December 2022 (UTC)[reply]

The meaning of "bot" may be unclear here. In the context of Wikipedia (per Wikipedia:Bot_policy#Definitions), a "bot" is a software program that edits autonomously without user input; there do not currently exist any language models capable of independently establishing API connections to Wikipedia and making edits without human interaction. If they did (this is a horrible idea) it would be covered under the bot policy and require a WP:BRFA. The policy under which BRFAs are required does not apply to assisted editing (i.e. the use of software to create letters, numbers and symbols that were not produced by a human being pressing a keyboard). This is governed by existing policies (such as WP:MEATBOT and by the guideline at WP:ASSISTED. jp×g 09:28, 13 December 2022 (UTC)[reply]
@JPxG: The entire news field refers to ChatGPT as a chatbot. It is general consensus that it is a bot. ChatGPT speeds up writing, by doing it for (that is, assisting) the user, which falls under the "higher speed" and "assisting human editors" foci of the bot policy. There is a passage in the bot policy that covers policy contradictions (such as between the lead and definitions sections), and situations where the spirit of the rule and its precise wording conflict, that is, cases of ambiguity. In its definition of "Bot Approvals Group" (BAG), the bot policy states: "The BAG also determine the classification as bot or assisted editing, in ambiguous cases." According to WP:ASSISTED, it is up to the Bot Approvals Group to decide whether bot approval is necessary. Based on the previous 2 sentences, BAG decides whether use of particular software falls under its jurisdiction. It remains to be seen what BAG's reaction(s) to LLMs, and the chatbots built upon them, will be.    — The Transhumanist   11:10, 14 December 2022 (UTC)[reply]
I do not think you are properly acquainted with how this software works: like I said, there do not currently exist any language models capable of independently establishing API connections to Wikipedia and making edits without human interaction. No media outlet has ever claimed that ChatGPT falls under the English Wikipedia's definition of an automatic bot – and even if they did, they do not determine policy. It is true that WP:MEATBOT and WP:ASSISTED are part of the bot policy, but there is a very clear definition of what a "Wikipedia bot" is, and it's defined by that same policy. At any rate, all edits (whether made by bots, software, humans using software, aliens using software, or Nagato Yuki psionically connecting to Wikimedia servers) are governed by existing policies and guidelines. To specifically address LLM output, a new policy would need to be written and ratified (which I am currently drafting a proposal for). jp×g 11:26, 14 December 2022 (UTC)[reply]
@JPxG: I believe the bot policy has wider jurisdiction than the narrow interpretation that you have presented. Establishing API connections is irrelevant, because a human is inserting bot-generated content. It's a bot-involved process. And those are encompassed by the bot policy which makes it up to BAG. A new policy could establish an exception, and I imagine the discussions will be extensive, as this is not a cut and dried case -- it is a sensitive issue with many potential ramifications. But, until such a policy is in place, this issue falls under BAG's jurisdiction, since they are the ones who decide the classification of a software program as it pertains to the bot policy.    — The Transhumanist   11:52, 14 December 2022 (UTC)[reply]
(edit conflict) WP:ASSISTED is deliberately vague to not restrict use of common scripts and tools. So it specifically says that only once volume becomes significant, such editing becomes more likely to be treated like a bot and BAG can determine this. It doesn't make it a bot, but it will be treated like a bot. We've never encountered any large-scale edits with LLM before, but we sure have seen a lot of high-volume editing. Half the bot policy only exists because of all the ways editors have inadvertently created issues with mass edits. So at that point, other parts of the policy start to matter, notably WP:CONTEXTBOT - which does not allow edits where context matters. I'm not saying copy-pasting LLM output is immediately covered by bot policy, nor does it matter whether anyone considers LLM to be a "bot". But bot policy will kick in once someone starts to make a lot of edits. And any new guideline will have to reconcile with this or we need to change bot policy to reconcile with LLMs. —  HELLKNOWZ  TALK 12:02, 14 December 2022 (UTC)[reply]

@JPxG: Another possible approach for vetting users for use of LLMs is via user group membership (aka "rights"). Here are our current user groups:

Code User group
AC Account creator
Ad Administrator
AP Autopatrolled
B Bureaucrat
Ch CheckUser
Co Confirmed
ECo Extended confirmed
EFH Edit filter helper
EFM Edit filter manager
EM Extended mover
EvCo Event coordinator
F File mover
IM Import
IAd Interface administrator
IP IPblock-exempt
MM Mass message senders
N New page reviewer
O Oversighter
Ro Rollbacker
Rs Researcher
Rv Pending changes reviewer
TE Template editor

These indicate membership in user groups (see: user access-levels). They pertain to who is granted access to various features of MediaWiki and its extensions. Theoretically, a user group could be created without being attached to a program function (that part could just be left blank?). For example, you could have a group called "LLM", with everyone in that group approved to use large language models in their editing. I don't know if this is doable, though.    — The Transhumanist   08:01, 13 December 2022 (UTC)[reply]

I do not think there is anything in our P&G that would directly prohibit use of content created by LLMs, nor do I think it would be a good idea to try to do so. All that is needed is to continue to hold individual editors responsible for all edits they make, including the copying of content from any source, whether from LLMs or other sources. We probably should add language in appropriate places reiterating that editors are reponsible for insuring that all content that they add, including anything produced by an LLM, meets our P&G. - Donald Albury 13:12, 13 December 2022 (UTC)[reply]
@Donald Albury: LLMs automate writing (edits). The rules are very clear on this: it falls under WP's bot policy, in the very first sentence.[1]   Therefore, it would require a new policy to allow use of LLMs without need for approval from the Bot Approvals Group (BAG).    — The Transhumanist   09:24, 14 December 2022 (UTC)[reply]
If someone uses an unapproved script or bot to edit WP, that is a violation of the bot policy, whether or not they use an LLM to generate any content being added. If someone uses an LLM to create text which they then copy into Wikipedia without using a an unapproved script or bot, that is not covered by the bot policy, but the user remains responsible for insuring that the content conforms with policy and guidelines. There is no point in banning content created by LLMs, as we already require that content be verifiable from reliable sources, and I doubt we will be accepting any content created by an LLM as a reliable source anytime soon. The danger is that LLMs may create potential content with citations to pseudo-sources, but we can go after users repeatedly adding such content to WP for abusing the policies on verifiability and reliable sources, without regard to whether such content came from an LLM. Donald Albury 13:48, 14 December 2022 (UTC)[reply]
I don't think it's plausible that LLMs are covered by the bot policy. If they were, grammar checkers, spell checkers, and machine translation would be "bots". Jahaza (talk) 19:49, 14 December 2022 (UTC)[reply]
@The Transhumanist: ChatGPT falls under Wikipedia:Bot policy, but per the definitions section it does not fall within that policy's definition of a bot. Rather, use of it would fall under the definition of "assisted or semi-automated editing", and the relevant policy section is Wikipedia:Bot policy#Assisted editing guidelines. The section doesn't aim to draw a 100% hard line, but my reading is that limited of ChatGPT for clean-up on a limited number of articles by a user in a limited closely-supervised way may be something users can do if they are trusted to apply their common sense. It is "Contributors intending to make a large number of assisted edits" who "are advised to first ensure that there is a clear consensus that such edits are desired." Limited use of ChatGPT to a lesser degree than would trigger this may currently be outside policy. In any event "A bot account should not be used for assisted editing".
It seems to me that an addition to the policy along the lines suggested by User:JPxG to address this potential hole might well be useful, eg "tools capable of assisting editors make substantial edits (for example large language model output) should only be used by competent editors who do not blindly paste tool output into the edit window and press "save"." Jheald (talk) 19:36, 18 December 2022 (UTC)[reply]
In what way does the bot policy cover ChatGPT? Just because it is called a "chatbot", doesn't mean it is a bot. Copying text from GPT-3 doesn't automatically become bot-like editing. Semi-automated edits? i'd call that borderline. It only becomes a problem (e.g. meatbot problems) if the amount of supervision needed to save an edit is below normal editing, and that the speed of the edits are above normal. (see awb, huggle, etc) 0xDeadbeef→∞ (talk to me) 10:20, 20 December 2022 (UTC)[reply]
The thing is, any LLM additions will inevitably be both faster than writing manually and, due to its confidently-wrong output, less reviewed. Otherwise, why would anyone bother with it? I feel that assuming that editors will spend just as much time to carefully review the LLM output is wishful thinking. I'd like to be proven wrong, but I have never seen any precedent on Wikipedia that better tools would lead editors to spend the time saved to further verify the tool output. If anything, tools only create induced demand. —  HELLKNOWZ  TALK 21:18, 20 December 2022 (UTC)[reply]
I don't think we need to do anything in particular. There is plenty of confidently-wrong content being added to Wikipedia by human editors already and we're dealing with that as well as we can. I think the intersection of 'Editors who will use a cutting-edge AI to generate content' and 'Editors who will do this without validating the output' is a very small overlap and will be of such small volume to be picked up by other editors as usual. A huge influx will be detected in aggregate, and we can deal with that if it becomes a problem in the future. If someone uses LLM to generate confidently-right content or articles, that's indistinguishable from content generated by a competent human, I refer you to xkcd: Constructive! A simple but unobtrusive first step may be to tag an edit as 'generated by AI', or maybe just ask editors to add a tag to their user pages if they regularly do so, but the intersection of problematic users who also follow this would be basically non-existent. JeffUK (talk) 10:40, 24 December 2022 (UTC)[reply]
@JeffUK, Hellknowz, 0xDeadbeef, Jheald, Jahaza, Donald Albury, JPxG, and AndyTheGrump:

So, wait until after it becomes a huge influx/problem, and only start to deal with it then? What if a solution takes weeks or months to develop?

By the way, what might the solution be for a huge influx of LLM-generated content, and how long would such a fix likely take?    — The Transhumanist   11:32, 2 January 2023 (UTC)[reply]
I am having trouble understanding what you are talking about at this point. I wrote WP:LLM some weeks ago, a gigantic proposal for a comprehensive guideline on the use of LLMs, and linked it multiple times on this noticeboard. While it is not complete, it seems to me like it covers everything you are talking about here. Do you have an opinion on it at all, or...? jp×g 15:39, 2 January 2023 (UTC)[reply]
@JPxG: I was responding to JeffUK's statement "A huge influx will be detected in aggregate, and we can deal with that if it becomes a problem in the future." Intently waiting until something becomes a huge problem before you start dealing with it, sounds like a disaster waiting to happen. Also, what good are guidelines going to do if the average person is using chatbots on a regular basis? People just jump in and edit Wikipedia without reading any project-level pages first. If there's a huge influx, and all you are doing is holding up a sign that says "Read this", what good will that do? You haven't addressed how the problems associated with a potential huge amount of chatbox input (in the form of one-off edits from a large number of people) would be prevented or processed. One solution is to fix the chatbots themselves, so that they don't generate Wikipedia-incompatible content in the first place, which would require working with the developers. A second method would be to create bots to detect and remove either chatbot-generated content, or if possible, policy-breaking content. Simply writing policy and hoping no flood comes, just doesn't seem like a viable approach should a flood hit. That approach may work for the first 3 or 4 years, but what if the flood comes in the 5th year and Wikipedia isn't prepared? We will have wasted 5 years that could have been spent preparing. Maybe we'll be lucky and chatbots will be smart enough to read and follow your guidelines. But if they are not? Fortunately, Doug Weller has passed word along to the Wikimedia Foundation. Maybe they will do something other than write editing guidelines.    — The Transhumanist   03:09, 10 January 2023 (UTC)[reply]
@JPxG although I haven't read more than the beginning, I'm also worried about AIs creating images.For instance I've seen some extremely convincing ones of fake archaeological sites and artefacts. Couldn't people pass them off as their own photos? Or am I missing something? Doug Weller talk 17:06, 2 January 2023 (UTC)[reply]
Yeah, that's a whole other deal. This proposal is only for large language models. Large image models will probably need to be governed by something much more imaginative. jp×g 17:11, 2 January 2023 (UTC)[reply]
If someone's editing is inappropriate, the solution will be notifying them it's inappropriate, warning them, then banning them if they don't stop. There are ways for incompetent editors to make massive plausible seeming changes to the encyclopaedia right now. e.g. by copy/pasting content from other places, or just writing in made up 'facts', LLM really won't make this any easier for someone who's intent on doing this. JeffUK 18:09, 2 January 2023 (UTC)[reply]
I'm kind of wondering what in the heck we're standing to gain by creating any sort of policy surrounding ChatGPT and its ilk. 🌈WaltCip-(talk) 13:24, 9 January 2023 (UTC)[reply]
If these AIs are used in some way for a large scale creation of articles, I think that will be a disincentive for a lot of editors and may drive some away. I disagree with JeffUK on the simplicity of dealing with this. First, you need to be able to spot them and that's work. Secondly, that also assumes that the numbers will be small. Doug Weller talk 15:48, 9 January 2023 (UTC)[reply]
@WaltCip: None. Most editors don't read policy before editing. That's because most editors post very infrequently. But there are a lot of them, and they have authored of most of Wikipedia. What happens when they are all using chatbots, much in the way that most everyone today uses Google?    — The Transhumanist   03:34, 10 January 2023 (UTC)[reply]
"Someone." Apparently, you are assuming it will be one person, or a small enough number to be handled manually. But, what if over the next few years chatbots become ubiquitous with almost everybody using them? How will you deal with it when half the content contributions to Wikipedia are being generated using chatbots?    — The Transhumanist   03:26, 10 January 2023 (UTC)[reply]

Develop software to detect it?

Someone seems to have done this, see A college student created an app that can tell whether AI wrote an essay Maybe the WMF should look into software detection of AI material? Doug Weller talk 14:45, 10 January 2023 (UTC)[reply]

We have mw:ORES that uses machine learning to detect vandalism, so the infrastructure is already in place. All we need to do now is to add the dataset. CactiStaccingCrane (talk) 16:06, 12 January 2023 (UTC)[reply]
(ORES is used for these "likely have problems" and "likely bad faith" highlights in Special:RecentChanges) CactiStaccingCrane (talk) 16:08, 12 January 2023 (UTC)[reply]
Even if ORES is up to the task, and it isn’t perfect now, you still need enough editors to deal with large numbers. Doug Weller talk 18:21, 12 January 2023 (UTC)[reply]
@Doug Weller, or a bot. — Qwerfjkltalk 21:02, 12 January 2023 (UTC)[reply]
@Qwerfjkl: What would the bot do?    — The Transhumanist   22:38, 12 January 2023 (UTC)[reply]
@The Transhumanist, revert additions and/or tag articles. — Qwerfjkltalk 07:03, 13 January 2023 (UTC)[reply]
@CactiStaccingCrane, Doug Weller, and Qwerfjkl: All we need to do is add what data set? You make it sound easy (keeping fingers crossed). What does that entail?    — The Transhumanist   22:38, 12 January 2023 (UTC)[reply]
OpenAI have annouced they are adding in some kind of lexical watermark than can be used to identify any output from ChatGPT. scope_creepTalk 13:08, 16 January 2023 (UTC)[reply]

Other inherent problems only partially touched on

Other inherent problems only partially touched on:

  • Editing articles involves also understanding what is already in the article and how it is organized plus understanding and interpreting policies and guidelines.
  • What's unspoken but runs through many things including current Wikipedia is sort of a commensurate investment. You can get volunteers to take their time to review and deal with issues because they know they are dealing with something that an editor has invested time in to create. Part of the reason that we don't allow mass creation of articles by bots. In other words, we'd significanly lose volunteer efforts
  • Modern AI is inherently unaccountable black boxes. There is no way to see or interrogate or demand/recieve accountability or reasoning for how it arrived at what it arrived at.
  • If gibberish or semi-gibberish is created, it normally requires an expert to spot and remove it..... a very scarce resource. I once uncovered a set of technical-subject articles (about 100 article as I recall) which looked very technical and Wikipedian and were sourced but if you knew the subject you knew were pure gibberish.

Sincerely, North8000 (talk) 22:23, 10 January 2023 (UTC)[reply]

Thanks. I agree entirely. Doug Weller talk 09:04, 11 January 2023 (UTC)[reply]
Well said. I think that folks are overestimating the ability of our review processes to detect "vaguely plausible bullshit" - it's not very common for human editors to fill in blanks with made-up facts and numbers, and I'm not sure that AfC or NPP are checking for this as it would greatly increase their workload. –dlthewave 19:42, 11 January 2023 (UTC)[reply]
[I]t's not very common for human editors to fill in blanks with made-up facts and numbers. Maybe not when adding content, but I see this happen all too often in edits to temperature tables in climate sections. Of course, the tell there is changing temperatures without citing a source or commenting about correcting from a cited source. - Donald Albury 20:17, 11 January 2023 (UTC)[reply]
As I recall, the big one that I caught looked like some type of expose project or research project to see if such a scam could get far in Wikipedia. It was sort of in mashup of words from actual sources. Total nonsense, but a typical reader might think it was simply over their head. North8000 (talk) 21:25, 12 January 2023 (UTC)[reply]

Summary of discussion so far

@Aquillion, Andrew Gray, Fram, Levivich, Ritchie333, 0xDeadbeef, ONUnicorn, JPxG, EpicPupper, Sojourner in the earth, Dlthewave, Doug Weller, Qwerfjkl, CactiStaccingCrane, WaltCip, JeffUK, Hellknowz, Zero0000, AndyTheGrump, Bluerasberry, David10244, Boud, Ziko, Pharos, Andrew Gray, WhatamIdoing, Tazerdadog, Barkeep49, Tigraan, Blueboar, MJL, PerfectSoundWhatever, Koziarke, SmallJarsWithGreenLabels, Isaacl, Lee Vilenski, Thebiguglyalien, Hanif Al Husaini, and Xeno:

Highlights of the discussion so far:

  • Chat-GPT is taking the world by storm (translation: it has gone viral).
  • Chat-GPT, and other LLM-based chatbots, can generate compositions, some good enough to pass as college-level essays.
  • Wikipedia is included in the corpus (training data) of Chat-GPT (and other chatbots).
  • Such software has the potential to be used for:
    • Generating Wikipedia content, including writing new articles and adding new material to existing articles.
    • Generating Wikipedia policy content.
    • Generating discussion content, such as on policy talk pages. That is, editors using it to write their discussion replies for them.
    • Editing articles, including rewrites, and using chatbots as a grammar checker.
    • Editing other namespace pages, such as policy pages, etc.
    • "Can be used for an extremely wide variety of tasks across the project, from formatting to table syntax to HTML generation to copyediting." (quoting JPxG)
    • Creating hoaxes with less effort.
  • Most Chat-GPT output lacks citations.
  • Some experiments were run, showing that Chat-GPT:
    • Copies writing styles very well.
    • Has a tendency to make things up, yet presents it as fact in an encyclopedic tone. One editor dubbed this "confident nonsense". In one experiment, Chat-GPT created an article reporting that Wikipedia's own Signpost newsletter was the recipient of several Pulitzer Prizes.
    • Can include references, but some of the references were made up and totally fictitious.
    • Some references cited Wikipedia (an ineligible source for Wikipedia articles).
    • One of the experiments generated instructional content, a recipe, that the user followed, and ate the results of.
    • Another experiment used Chat-GPT to answer hypothetical questions in the style of WP's teahouse department. It worked fairly well.
    • Yet another experiment created a sample policy page, showing that chatbots are not limited to editing articles. They can generate or edit pretty much any type of page on Wikipedia, except files (images).
    • Chat-GPT output is not fact-checked.
    • Chat bots don't actually understand what they are writing.
    • When used responsibly as a tool, with editors carefully prompting the chatbot, and editing and fact checking its output before posting it to Wikipedia, a chatbot can be very useful and increase editor productivity: the LLM GPT-3 was successfully used to create department reports for Wikipedia's newsletter, The Signpost.
    • JPxG conducted an experiment/demonstration to show that Chat-GPT is a sophisticated interactive editing tool, which you tell it what you want it to do to a textual work, and then it does it. See it here: User:JPxG/LLM demonstration.
  • It was pointed out that Wikipedia policy already covers all contributions, whether generated by chatbot or human. Ultimately, the user is responsible for material they copy and paste into Wikipedia.
  • Issues of concern that were raised include:
    • Users copying chatbot-generated text into Wikipedia without carefully editing and fact-checking it first.
    • Confident nonsense (misinformation generated by chatbot) may be hard to spot.
    • The potential of chatbots to violate copyright, by directly copying, or generating text based on, copyrighted works.
    • Violating Wikipedia's licenses, most notably the attribution requirements. Chat-GPT output generally does not include attributions.
    • A chatbot-edited Wikipedia could wind up in the training data for those same chatbots (or their next versions), creating a potentially error-compounding feedback loop.
    • The suggestion was made to prepare for a potentially large future increase in chatbot entries to Wikipedia, by:
      • Working with chatbot developers to make chatbot-generated output Wikipedia compatible.
      • Develop bots to identify and process chatbot entries.
  • No consensus has emerged on what the Wikipedia community should do about LLMs/chatbots. Some editors think that policies/guidelines and the current editor pool could handle any influx of chatbot generated edits. Some other users were concerned that there is potential for LLM/chatbot contributions, such as one-off edits by members of the general population, to overwhelm our pool of editors. One user pointed out that it may take experts to discern nonsense articles, and experts on Wikipedia are a scarce resource.
  • Consensus did emerge on something not to do. It was agreed that banning chatbot-generated content was not a good idea at this time, and probably wouldn't work anyways.
  • Software has been developed to identify Chat-GPT-generated text.
  • It appears some editors may take the initiative to prepare for a worst-case scenario (chatbot input going beyond our editor pool's ability to handle), and discussion on how to do this has begun.
    • WP:ORES could theoretically be trained to identify chatbot edits.
  • The Wikimedia Foundation has been contacted about the concern over LLMs/chatbots, presenting a contact there with a link to this and a previous discussion.

Did I miss anything?    — The Transhumanist   01:22, 13 January 2023 (UTC)[reply]

AI-generated images are rapidly becoming a Big Thing, so it is not correct to exclude them. Also, "Wikipedia policy already covers all contributions, whether generated by chatbot or human" is misleading as it is true only by accident. A more precise description would be "Wikipedia policy was written without any consideration of chatbots". Zerotalk 03:28, 13 January 2023 (UTC)[reply]
I'm concerned about "Such software has the potential to be used for: creating content farms that good-faith human editors, including existing experienced editors, will sometimes mistake for reliable sources when they are writing content".
Also, the statement that "Software has been developed to identify Chat-GPT-generated text" is true, but not relevant for very short contributions. Some of this is using sentence length, and you won't be able to identify an abnormal sentence length if you only look at two or three sentences. WhatamIdoing (talk) 05:33, 13 January 2023 (UTC)[reply]
Anyway, detection methods that work now won't work in the next generation. Eventually (and not far in the future) distinguishing between human-written and computer-written prose will be impossible for practical purposes. This is going to be the greatest threat to Wikipedia since its founding. Zerotalk 05:51, 13 January 2023 (UTC)[reply]
I have started Category:Wikipedia essays about artificial intelligence, Perhaps folks here would like to add to the collection, and document yet more thoroughly! Pharos (talk) 01:07, 14 January 2023 (UTC)[reply]
Adjacent to hoaxes there's also the likelihood of spammers using GPT to bulk out their edits. I strongly suspect that the text of this edit today, writing repetitively about a static sculpture as if it was a functional scientific instrument, was generated with GPT-3, probably giving it a prompt to explain Orbital Reflector in terms of dark matter and black holes, the subject of the two embedded spam links. Belbury (talk) 11:39, 17 January 2023 (UTC)[reply]

Chat-GPT spreading fast

The situation is changing rapidly:

Chat-GPT may become ubiquitous sooner than previously thought, and so far, identification methods have fallen flat...

Here's some recent news:

  1. ChatGPT Will Be Everywhere in 2023 (CNET)
  2. Microsoft is reportedly integrating ChatGPT's technology into Bing (Yahoo)
  3. Microsoft is looking at OpenAI’s GPT for Word, Outlook, and PowerPoint (The Verge)
  4. There's a Problem With That App That Detects GPT-Written Text: It's Not Very Accurate (Futurism.com)

With the user base for Chat-GPT about to explode, the potential for Chat-GPT-generated text being added to Wikipedia will explode right along with it. It's looking uncertain whether or not Wikipedia's editor community will be able to keep up with the influx. In light of recent events, what should be done about this?    — The Transhumanist   03:21, 13 January 2023 (UTC)[reply]

As well as being able to write plausible-looking prose on any subject, computers can also be programmed to add it to Wikipedia all by themselves. The first task is to absolutely ban computers from editing, with the sole exception of authorized bots. The second task is to add to appropriate policy pages that all content (authorized bots excepted) must be added by a human and that that human is responsible for checking policy conformance of the content. Zerotalk 08:06, 13 January 2023 (UTC)[reply]
I couldn’t agree more. Does anyone have objections? Doug Weller talk 10:45, 14 January 2023 (UTC)[reply]
I was hoping to get more feedback on WP:LLM from having posted it here, but either way, I think it is pretty close to ready for consideration as a guideline (or policy, as appropriate)... based on the conversations I've had (and seen) I am prepared to write an RfC for its adoption. jp×g 11:00, 14 January 2023 (UTC)[reply]
@JPxG: The guideline is not ready. It is not where near complete, and it needs a rewrite. Here are some proofreading notes:

It's way too redundant, repeating policies and itself, without explaining how to get the job done. Aside from the "fit for" sections, the rest of the page can be reduced to a single paragraph.

It presents -- should only be used by competent editors who do not indiscriminately paste LLM output into the edit window and press "save" -- four times! Someone who is incompetent isn't going to be able to judge whether or not they are. Also, "indiscriminately" is vague. That entire sentence should be removed.

Editors need to know what they need to do to the text before they can press "save". For example, you alluded to a manner of using LLMs in compliance with WP copyright policy, but you didn't explain how. How can an editor be sure that an LLM-generated piece doesn't violate someone's copyrights? What's the procedure?

Rather than covering "good fit" and "not good fit", the guideline should present explicit instructions: "Use it for this" and "Do not use it for this". And then explain how.

I hope you find these observations and comments helpful. Sincerely,    — The Transhumanist   08:04, 16 January 2023 (UTC)[reply]
I agree that the page is not finished, but I don't really know what you are objecting to here. It kind of sounds like you are inventing problems – if users don't know how to check if things are true before putting them into Wikipedia articles, they shouldn't be editing at all. If users don't understand what copyrighted material is, they need to read Wikipedia:Copyright policy, which is linked to from this page when it's mentioned. That is an explanation of how to get the job done. It should not be necessary to create an exact copy of Wikipedia:Verifiability that says "When using a LLM," at the beginning of every sentence. jp×g 08:29, 16 January 2023 (UTC)[reply]
Like DALL-E last year, or NFTs the year before that. I'll believe it when I see it, and I can't see the value in spending even more time discussing a hypothetical future threat to Wikipedia. – Joe (talk) 08:39, 13 January 2023 (UTC)[reply]
The scariest yet most plausible thing is that this is happening with some of the articles but we aren't aware of it. I don't think raising awareness on this issue is a bad thing given how fast AI advances nowadays. CactiStaccingCrane (talk) 00:39, 14 January 2023 (UTC)[reply]
Sorry, I missed the "recent events". Where is the evidence for GPT problems on Wikipedia? —Kusma (talk) 11:57, 13 January 2023 (UTC)[reply]
Where's the evidence that Wikipedia can't cope with AI generated articles? doktorb wordsdeeds 14:07, 14 January 2023 (UTC)[reply]
@Doktorbuk: You are talking in terms of hindsight (asking to see what has already happened), rather than applying foresight to assess a potential threat by asking "What could happen?"

Here's an article from the New York Times -- imagine a similar effort directed at Wikipedia using thousands upon thousands of (seasoned) new accounts to support political POVs, revise history, censor opposing opinions, and spread other forms of misinformation:

https://www.nytimes.com/2023/01/15/opinion/ai-chatgpt-lobbying-democracy.html

It's only a matter of time before the powers that be shift their attention, and their tools, upon the English Wikipedia. The question is, are we ready for when we have to be? Here's an article that makes one wonder what these people will do now that they have Chat-GPT to work with:

https://www.theweek.in/news/world/2023/01/06/saudi-arabia-infiltrated-wikipedia-and-this-is-how-they-did-it.html

So, do we really need evidence that the English Wikipedia has already been breached by LLM-assisted POVers before proceeding? Or can we prepare for this in advance?    — The Transhumanist   00:34, 16 January 2023 (UTC)[reply]
Your reasoning seems to be that
  1. ChatGPT (or its equivalents) can write disinformation quickly and cheaply
  2. POV-pushers (governments, lobbies etc.) are currently limited by the time humans need to write disinformation
  3. Wikipedia is a prime target for such POV-pushers
  4. Therefore, ChatGPT (or its equivalents) will flood the gates, unless we do something.
I will grant you (1) is either already true or will likely be in the near future.
However, (2) is questionable (see that XKCD about old-fashioned human-crafted POV-pushing). I would guess coordinating the messaging and maintaining the disinformation is a much larger fraction of the costs than actually writing the text.
(3) is also dubious. Editing in a way that sticks is much harder on Wikipedia than in other places (such as facebook, reddit, etc.). Maybe it has more impact, but the cost-benefit analysis is not obvious.
Finally, inaction is always an option. It might not be a good option, it might even be the worst option, but it must be compared to other specific measures. "Something must be done" without specifics is just the politician's fallacy. In the absence of details about the threat, it’s hard to compare the possible countermeasures. TigraanClick here for my talk page ("private" contact) 16:41, 16 January 2023 (UTC)[reply]


Zero0000's post is a good start. A simple way to crystalize the situation is to ask the human editor for their rationale for a particular phrase. North8000 (talk) 03:55, 14 January 2023 (UTC)[reply]

I feel like additions of large, overly-verbose unsourced text are something of a giveaway. See, for example, the first revision of Artwork title, written by ChatGPT. — Qwerfjkltalk 11:06, 14 January 2023 (UTC)[reply]
You can be confident that any giveaways are on the chatbot writer's list of things to fix in the next generation. Zerotalk 11:56, 14 January 2023 (UTC)[reply]
They should also fix the problem of the chatbots making stuff up. Someone should ask the chatbot writers to turn off the poetry and fiction generation algorithms, and any other algorithms that make things up, when the chatbots are composing expository text. Or add new algorithms to handle expository writing. Just the facts. And sources.    — The Transhumanist   00:42, 16 January 2023 (UTC)[reply]

Is someone clear on what the copyright status of texts produced by LLMs is? From what I get, they may be considered derivative works from the dataset they were trained on. From [2]: As a result of the human authorship standard, “under U.S. current law, an AI-created work is likely either (1) a public domain work immediately upon creation and without a copyright owner capable of asserting rights or (2) a derivative work of the materials the AI tool was exposed to during training,” Esquenet continues. “Who owns the rights in such a derivative would likely be dependent on various issues, including where the dataset for training the AI tool originated, who, if anyone, owns the training dataset (or its individual components), and the level of similarity between any particular work in the training set and the AI work.” If they are derivative works then they cannot be published on Wikipedia just like this. Do we have more information on this? For example, does OpenAI specify somewhere the copyright status of the text produced by ChatGPT? Phlsph7 (talk) 09:09, 16 January 2023 (UTC)[reply]

The first question is whether a generated text that closely resembles an item from the training set is copyright infringement of that item. For instance, Microsoft Copilot happily outputs the Fast inverse square root code. I would expect that courts will judge such things to be copyright infringement. Copyright infringement statutes do not require to prove that the infringer copied a specific source (that would be difficult to prove), just that the content is substantially similar. Therefore, whether the tool is a simple ctrl-C ctrl-V or a sophisticated machine learning model should not make much difference.
The second question is whether OpenAI (or any other AI tool provider) can assert copyright on whatever the tools they provide create. The OpenAI terms of use seem relatively permissive, but others might be less generous. I do not know the answer to that question. I would hope they cannot, since they only provide tools (Microsoft should not be able to assert copyright on the text I write using Word, or the images I draw using Paint).
The third is whether a human using ChatGPT can assert copyright on ChatGPT answers, or otherwise constrain the use of the resulting text. The quote you give is probably based on the US copyright office’s position (taken during the monkey selfie copyright dispute): Because copyright law is limited to 'original intellectual conceptions of the author', the [copyright] office will refuse to register a claim if it determines that a human being did not create the work. However, giving a prompt to ChatGPT might or might not constitute significant creative input. The position that anything edited by a machine becomes public-domain is untenable (if I use an orthographic corrector on the draft of my novel, it does not turn it into PD), so it must be a question of degree. Also, non-US courts might have different opinions. TigraanClick here for my talk page ("private" contact) 16:06, 16 January 2023 (UTC)[reply]
As I understand it, the degree of access by the alleged infringer to the source text in question is a factor in determining infringement. Only a specific expression is protected by copyright; if you and I independently write the same sentence, one is not a copyright violation of the other. The amount of similar text also plays a role, since the larger it is, it's more improbable that it was created without copying.
Facts and natural laws can't be copyrighted; this also covers algorithms (though a particular expression can be copyrighted). So I can't copyright a single instance of a Javascript for-loop and claim rights to all Javascript for-loops as derivative work. In cases where the learning model creator is explicitly providing its model for use as a tool, I think (disclaimer: not a legal opinion) it is reasonable for this to be the same as a work for hire. Thus if the result is eligible for a new copyright owner independent of any source texts, the tool user would be the owner. (If I use a spellchecker on the latest bestselling novel, the result is not eligible for a new copyright owner.)
To be really safe, we'd want language models trained on public domain text. But I think it could be argued with a really large model trained on, say (just drawing numbers out of air), hundreds of thousands of documents with thousands of independent authors, the resulting correlations can no longer be attributed to specific input text, for cases where the output is not a significantly long passage substantially similar to a specific source text. isaacl (talk) 18:05, 16 January 2023 (UTC)[reply]
One of the main issues to deal with would be the following: an editor tells ChatGPT to write a text on a topic and then adds this text in the form of a paragraph/section/article to Wikipedia and thereby publishes it under Creative Commons/GNU license. The question is: what are the chances that this constitutes some form of copyright violation? This might concern specifically problems with the 1st and the 2nd question addressed by Tigraan, i.e. whether the copyright of someone whose work was part of the training set was violated and whether openAI's copyright was violated. For the first question, it's probably relevant what the copyright status of the texts in the training set is and how similar the produced text is to the texts in the training set, as isaacl points out. Answering these questions would be quite relevant for any Wikipedia policy on the topic, like the one JPxG is currently drafting. Phlsph7 (talk) 07:18, 17 January 2023 (UTC)[reply]

Acupuncture

A case request was opened about 24 hours ago at DRN concerning acupuncture, and I closed it for various reasons, one of which is that there is a consensus in Wikipedia that acupuncture is not medically effective. The filing editor is continuing to ask questions, and so I will ask questions here. Just to be clear, I am a skeptic about forms of so-called alternative medicine, including acupuncture, and I know that Wikipedia is skeptical about so-called alternative medicine, but I would like to be directed to where that consensus has been established.

So, first, will someone please point me to where the consensus has been established that Wikipedia considers acupuncture to be pseudoscience?

Second, if an editor wants to challenge the existing Wikipedia consensus that acupuncture is pseudoscience, what is the correct forum for that purpose?

Thank you. Robert McClenon (talk) 20:15, 19 December 2022 (UTC)[reply]

Moved to WP:FTN
Any consensus arrived at via an RfC at a WikiProject would represent WP:Advice page perspective at most: it cannot be treated as standing consensus for a standard approach to be followed in individual articles, except when a WP:LOCALCONSENSUS at the relevant article talk page adopts the same perspective directly, or once the community adopts the rule into WP:Policy through a WP:PROPOSAL or if involves language added to the WP:MoS through the normal process. So I would caution that Robert should probably not direct the editor in question to look in that direction, or at least add a lot of caveats if they do. SnowRise let's rap 01:07, 9 January 2023 (UTC)[reply]
  • @Robert McClenon, this question is actually more complicated than it sounds. Here's an oversimplified answer:
  • Acupuncture, including placebo/sham acupuncture, often makes people feel better, especially for conditions that are susceptible to placebo effects (e.g., pain).
  • The idea that there is a non-material energy force that goes through certain paths/points in human bodies, and that the act of sticking a needle in those spots changes the flow of that energy is not compatible with scientific knowledge. Scientific knowledge deals with the material world; anything that is not part of the material world is non-scientific (e.g., beauty, goodness, spirituality, knowledge per se, etc.).
  • Something can "work" (e.g., function to some purpose; be accurate) and be non-scientific (e.g., all great works of art; all good journalism). Something can "not work" and be 100% scientific (e.g., broken engineering designs, experimental drugs that turn out to be ineffective or harmful).
  • What makes something pseudoscientific is the story you tell about it. If I tell you that a television works because miniature actors run around inside of it to make the picture, then: the television works and my story is nonsense. If I tell you that a television works because string theory says that dark matter interacts with weak forces to make the picture, then: the television works and my story is pseudoscience. If I tell you that a television works because it receives radio waves that tell it which bits of the screen to turn off and on, then: the television works and my story is scientific.
  • Acupuncture could be said to work (for some conditions, under some circumstances), but the story about chi is irredeemably pseudoscientific.
Is it fair to say that acupuncture is pseudoscience? Maybe. It's an incomplete explanation, but IMO it might be fair.
See also Wikipedia:Alternative medicine and Wikipedia:Biomedical information if you want more general information. WhatamIdoing (talk) 22:27, 5 January 2023 (UTC)[reply]
  • @Robert McClenon: As far as I know, there is no such general consensus that operates at the overall community level to find that Acupuncture is pseudoscience and should be described as such, and I'd actually argue it was not appropriate to describe it as such to the editor in question. As Marshall noted above, the topic is under a WP:DS advisory, but that's a behavioural rather than a content matter. Any consensus on this topic one would expect to find on the talk page, in the TP archives, and in the edit revision history for Acupuncture and related articles. That is to say, the WP:LOCALCONSENSUS for those questions. WikiProjects also might provide insight to how editors working in this area feel about such issues, but per my response to Marshall above, those views are WP:Advice pages perspectives and not binding consensus.
It sounds like you had other good reasons for closing the request (and you would know what DRN's requirements are if anyone), and it probably is fair to say that consensus on this issue is unlikely to change, but that's not because of any established Project-wide stance that can be pointed to outside of the talk page discussion on the relevant pages. Rather it is because the belief in many of the claims made by acupuncture proponents cannot be supported by scientific literature to the WP:MEDRS standard, such that WP:V and WP:WEIGHT can be satisfied. That and a healthy dollop of skepticism from the typical Wikipedia editor towards alternative medicine claims, exacerbated by the fact that most veteran editors will have had to deal with POV pushing by SPA proponents of various practices, which probably does heighten the bias a little, if it can be fairly called that.
But there's no reason, in principle, that consensus can't change at the relevant articles (indeed, it's established policy that it always can), and that is where you should direct your editor: to the individual talk pages. Just be sure to remind them that arguing the point against recent and/or very strong consensus on the issues can be found to be WP:Tendentious after a point. That said, WhatamIdoing's caveats above are worth remembering: some claims about acupuncture's physiological effects are more supported by evidence than others, and the medical establishment is hardly 100% aligned against it, because one can still observe a limited effect in narrow circumstances on the one hand, and yet also dismiss other more wild claims of efficacy and what it can treat or accomplish, as well as any related mysticism or snakeoil peddling. SnowRise let's rap 01:27, 9 January 2023 (UTC)[reply]
Thank you. I think that the issue has gone away, because the editor who was causing the issue has gone away or quieted down. As I mentioned, there was a request at DRN which I closed. The filing party was an unregistered editor who had been partially blocked from the Acupuncture talk page for bludgeoning a discussion, and who wanted to contest their partial block, and who said that our criticism of Acupuncture was libelous to acupuncturists. DRN is not a forum for block review, or for wild claims that articles are libelous. I think that the issue has blown away. Robert McClenon (talk) 21:47, 9 January 2023 (UTC)[reply]

Separating Fact from Fiction In Movie and TV descriptions

I was extremely disappointed with Wikipedia's posting regarding the feature film Tar. Your listing failed to mention that the Protagonist is not a real person. The response I received from my initial contact with Wikipedia concerning my initial complaint was "Must every article about a movie about a fictional character specifically state that the character is fictitious? Since most movies are not biopics, but made-up stories, doesn't it make more sense to highlight the few rather than the many? Tár is described as "a 2022 psychological drama film." See, for example, <https://en.wikipedia.org/wiki/Young_Winston>, described as "a biographical adventure drama war film covering the early years of British Prime Minister Winston Churchill."

My reply was, yes, it should be stated when it is a fictional character. In the case of Tar, Wikipedia merely bolstered the fake ad campaign put out by the studio and/or producers that the film was a biopic about a real person. Wikipedia could have simply added to their page, as they did with 'This is Spinal Tap', "This Is Spinal Tap is a 1984 American mockumentary film...". Fisher92F (talk) 18:12, 28 December 2022 (UTC)[reply]

Usually we presume all characters in a work are fictional, but the way the Tar page is written, particularly the plot, it definitely raises questions about Tar being real or not (the number of real life touch points on the plot). Thus is definitely a case that "fictional" needs to be added to the lede and plot section. Masem (t) 18:26, 28 December 2022 (UTC)[reply]
That Lydia Tár isn't linked in the article (along with the film being labeled a psychological drama) should clue the reader in this is a fictional story. I haven't seen the film, but unless its presented as documentary the sort of mockumentary label used on This Is Spinal Tap wouldn't make sense. (That This Is Spinal Tap was a fictional documentary was central to the film's premise.) That said, I added a sentence to the lead about viewer confusion about the fictional nature of the film and it looks like there are a number of other sources discussing this, so it could probably be expanded upon as a section in the body of the article. — Carter (Tcr25) (talk) 18:51, 28 December 2022 (UTC)[reply]
I've tagged it with the standard {{in-universe}} notice. This happens all the time in fandom-related articles. Boba Fett is more real to some people than Jesus Christ. The "Reception" section doesn't yet seem to address the marketing aspect. I noticed one footnote that they've actually included a "thinly veiled" real person as a character. Elizium23 (talk) 08:07, 2 January 2023 (UTC)[reply]
User:Fisher92F I don't know who you contacted, but the right place to discuss the article content is on the article's talk page here Talk:Tár, where more interested editors will see your query. To the above points, Forrest Gump was written as if it was a real story, interspersed with real events, we don't call Gump 'fictional' explicitly in the lead and I don't think we need to. JeffUK 10:19, 3 January 2023 (UTC)[reply]

Acceptable source?

I own a few of my grandfather’s sports medals and would like to mention given achievements in his Wikipedia article, but i wonder if it works as a source? I understand it possibly being problematic policy wise, but they’re not written about anywere (as sometimes is the case with these), so the idea of nobody ever knowing about it is tragic. Wonder if there are any exceptions that can be made. Xәkim (talk) 08:56, 1 January 2023 (UTC)[reply]

The medals themselves aren't really a usable source because nobody else can verify them. If your grandfather's achievements were reported somewhere (e.g. in local newspapers) that would be useable. Caeciliusinhorto (talk) 09:52, 1 January 2023 (UTC)[reply]
Remember that you can still create a web page descibing your grandfather's achievements, even if the medals cannot be used as a reliable source for a Wikipedia page. Wikipedia may be very popular, but its policies do not apply to the whole Internet. Phil Bridger (talk) 10:00, 1 January 2023 (UTC)[reply]
If I recall correctly, there is a template for citing a highway sign, so there are pretty wide boundaries on what is considered verifiable, although the common thread is that it's publicly verifiable. jp×g 16:51, 1 January 2023 (UTC)[reply]
{{cite sign}} is only intended for existing (as of the time of citation) public displays. Historical signage could be cited through a secondary source, including via archives. 63.118.97.194 (talk) 15:35, 2 January 2023 (UTC)[reply]
Your grandfathers article is interesting. Could you not photograph the medals and show them in the article and thus list the achievements? CambridgeBayWeather, Uqaqtuq (talk), Huliva 01:19, 2 January 2023 (UTC)[reply]
Well, CambridgeBayWeather can i do that? Xәkim (talk) 06:11, 2 January 2023 (UTC)[reply]
Or, maybe i can upload them to Flickr and give that as a a source? Xәkim (talk) 06:15, 2 January 2023 (UTC)[reply]
I can't give you a definitive answer because I don't know how the laws would cover reproductions of that sort in your country. CambridgeBayWeather, Uqaqtuq (talk), Huliva 14:33, 2 January 2023 (UTC)[reply]
Including photos of existing medals would be an example of WP:SKYISBLUE. If you want guidance on best practices for including medals in an article, WP:MILHIST WikiProject may have a lot of practical experience to share, given that it's a medal galore there. ~ 🦝 Shushugah (he/him • talk) 21:34, 2 January 2023 (UTC)[reply]
WP:RSN may be a more appropriate venue for this question. In any case, I can't imagine that most Wikipedia editors would accept the word of another Wikipedia editor, particularly one who is also claiming a clear conflict of interest, as a reliable source. It's not personal, we just need a source that has been published, preferable one that is secondary and independent. ElKevbo (talk) 03:36, 6 January 2023 (UTC)[reply]

Hi all,

Apologies if this is the wrong place for this, but whatevs.

I have recently added the IMDb profile of Barton Heyman to that article following the WP:IMDB-EL "explanatory essay".

The problem is, that IMDb page appears to include text that would appear to be a copyright violation of the 1996 NYT article I cited as reference in the article.

It would appear to me that WP:IMDB-EL does not take this sort of thing into account. Your thoughts about this?

Peter in Australia aka User:Shirt58 (talk) 🦘 11:53, 3 January 2023 (UTC)[reply]

If you believe that the content on the IMDB page may violate a copyright, then WP:COPYVIOEL applies; i.e., we cannot have a link in the article to that page. - Donald Albury 14:52, 3 January 2023 (UTC)[reply]

I think User:Nfhgfjtutumtmrm is not a sockpuppet of user:kfjisee

The following discussion is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.


example editing

  1. However, the reduction of delayed onset muscle soreness may require compression in combination therapy.]
  2. However, some people who are thrifty and don't buy products. change in new technology It may not affect your liking.]
  3. remove "to benefit people, enterprises and the planet. text...because it is like an advertisement and reference Not certifying that it's really good.
  4. The page you're looking for doesn't exist.--->http://www.caa.co.za/resource%20center/AIC'S/17.1.pdf CAA Website,no that link is show text with "CAA Website" and link to "http://www.caa.co.za/resource%20center/AIC'S/17.1.pdf"]
  5. not warmimg

101.109.178.43 (talk) 07:56, 4 January 2023 (UTC)[reply]

What? I think you are in the wrong place -- this sounds like an WP:ANI issue, although I doubt that anyone there will know what to do about warmimgs such as this. jp×g 11:22, 4 January 2023 (UTC)[reply]
this is​ ​not​ sockpuppet​ issue, please read topic again— Preceding unsigned comment added by 2001:44C8:470E:93F5:EA48:C3BE:D2E:C342 (talk) 00:09 6 January 2023 (UTC)
WP:ANI is the correct venue for reporting problem editing. However, I think you need to more clearly explain what the problem is in order for anyone at ANI to comment on it. - Donald Albury 14:35, 6 January 2023 (UTC)[reply]
Note, this page is only for discussing already proposed policies and guidelines and to discuss changes to existing policies and guidelines.xaosflux Talk 14:51, 6 January 2023 (UTC)[reply]
The discussion above is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.

Removing blocks made with the summary "Appeal is only to the Arbitration Committee" from the Admin Policy

The following discussion is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.


Should the Administrator and Blocking policies be harmonized by removing the bulletpoint that begins "Blocks made with the summary "Appeal is only to the Arbitration Committee" " from the Administrator policy? 21:46, 6 January 2023 (UTC)

Proposed Special situations text

In some situations, the usual policy for reversing another administrator's action does not apply:

* Blocks made with the summary "Appeal is only to the Arbitration Committee": Rarely, in blocking an editor, an administrator will have to note that their block "should be lifted only by the Arbitration Committee" or that "any appeal from this block is to ArbCom only". Such a provision must only be made if the nature of the block demands that its circumstances not be further discussed on-wiki (and instead be considered further only in a confidential environment). This could include situations where discussion would reveal or emphasize information whose disclosure could jeopardize an editor's physical or mental well-being, where on-wiki discussion would identify an anonymous editor, or where the underlying block reason would be defamatory if the block was unjustified. In such cases, the blocking administrator should immediately notify the Arbitration Committee by email of the block and the reasons for it.

In August 2012, the Arbitration Committee issued a reminder that administrators must promptly notify the committee when making sensitive blocks or when noting that a block can be "appealed only to ArbCom". In these situations, the administrator retains responsibility for their block (see this arbitration ruling) but will be accountable to the committee. (Such blocks have been the subject of long-standing Wikipedia practice, and were also discussed in the fourth paragraph of this statement.)

  • Blocks made by the Arbitration Committee: Separate from the first situation, a member of the Arbitration Committee may block an account. Blocks made by an arbitrator with the summary "For the Arbitration Committee", "Appeal is only to the Arbitration Committee", or "{{ArbComBlock}}" are made on behalf of the Arbitration Committee. These blocks are made by a decision of arbitrators, very rarely, and only with good reason. Therefore, administrators must not reverse ArbCom blocks without the prior, written consent of the committee. (See also: Wikipedia:Arbitration/Policy#Appeal of decisions.)
  • Checkuser blocks: Blocks designated as "Checkuser blocks" (that is blocks relying on confidential checkuser findings) may not be reversed by administrators who do not have access to the checkuser permission. Appeal of these blocks may be made to the Unblock Ticket Requests System (which has a designated "checkuser" area) or to the Arbitration Committee. Administrators were reminded in July 2010 that they may not reverse checkuser blocks without prior consent from the committee or a checkuser.
  • Oversight blocks: Blocks designated as "Oversight blocks" (that is blocks relying on information that has been suppressed) may not be reversed by administrators who do not have access to the oversight permission. The Arbitration Committee ruled in March 2013 that oversight blocks cannot be reversed without prior consent from the committee or an oversighter.

Background

There is longstanding wording in the Administrator's policy that allows Admins to make blocks "appealable only to the Arbitration Committee". There is no such authorization in the Blocking policy. The language in the Admin policy originated in a 2012 Arbitration Committee statement and references the need at the time to block editors for child protection and other oversightable activities. This statement came before the existence of Oversight Blocks and before the Wikimedia Foundation assumed responsibility for child protection enforcement. A 2022 review by ArbCom of the use of blocks labeled "Appeal is only to the Arbitration Committee" over the preceding 6 years revealed 11 blocks labeled like that not levied by the Arbitration Committee or by a Checkuser. Of those 11, 5 blocks were reported and 6 were not reported to the Arbitration Committee. Nearly all the blocks were for paid editing reasons. Following this audit the Arbitration Committee updated its guidance and the community subsequently affirmed that admins should not be blocking based on off-wiki evidence in an RfC.

Survey (ArbCom appeal blocks)

Discussion (ArbCom appeal blocks)

  • While brainstorming this idea Xaosflux suggested a larger change: use the same language in the Blocking and Admin policies. While that does seem useful, it's a much larger change than what I was aiming for. So I thought it worth bringing up in this discussion, should there be interest, without it complicating the more focused change I'm invested in. Best, Barkeep49 (talk) 21:46, 6 January 2023 (UTC)[reply]
  • If this is open, lets fix bullet 2 as well:
    FROM:
    Blocks made by the Arbitration Committee: Separate from the first situation, a member of the Arbitration Committee may block an account. Blocks made by an arbitrator with the summary "For the Arbitration Committee", "Appeal is only to the Arbitration Committee", or "{{ArbComBlock}}" are made on behalf of the Arbitration Committee. These blocks are made by a decision of arbitrators, very rarely, and only with good reason. Therefore, administrators must not reverse ArbCom blocks without the prior, written consent of the committee. (See also: Wikipedia:Arbitration/Policy#Appeal of decisions.)
    TO:
    Blocks made per the Arbitration Committee: The Arbitration Committee may require an account to be blocked as a remedy or motion. Such blocks must be labeled as "For the Arbitration Committee", "Appeal is only to the Arbitration Committee", or "{{ArbComBlock}}". These blocks are made by a decision of arbitrators, very rarely, and only with good reason. Only blocks approved by such as decision may be labeled as such. Therefore, administrators must not reverse ArbCom blocks without the prior, written consent of the committee. (See also: Wikipedia:Arbitration/Policy#Appeal of decisions.)
    Reasoning is that: 1, there no longer would be a "first situation". 2) There is no requirement to be an administrator to be on arbcom; arbcom must order these - but they are enacted by an admin - even if that admin just happens to be on the committee. (there is no "super class" of editors called arbadmins....) — xaosflux Talk 23:32, 6 January 2023 (UTC)[reply]
    @Barkeep49: any concerns? Wordsmithing, etc encouraged. — xaosflux Talk 23:34, 6 January 2023 (UTC)[reply]
    The entire paragraph is a bit longwinded. It could be half the size and still achive the same thing. For example: very rarely, and only with good reason why is policy making note of frequency and arbitrator motivations in a "don't unblock section"? All that needs to be said is "don't revert" and properly define an "arb block" but maybe not with those exact words... Terasail[✉️] 23:51, 6 January 2023 (UTC)[reply]
    I agree with Terasail that the section can be trimmed down to just its essentials, such as something like the following:
    • Blocks made by the Arbitration Committee: Blocks authorized by the Arbitration Committee must have a summary containing "For the Arbitration Committee", "Appeal is only to the Arbitration Committee", or "{{ArbComBlock}}". Administrators must not reverse such blocks without the prior, written consent of the committee. (See also: Wikipedia:Arbitration/Policy#Appeal of decisions.)
    The context of these blocks can be described within the arbitration policy and procedures. isaacl (talk) 00:06, 7 January 2023 (UTC)[reply]
  • @Barkeep49: This might just be me but what is with this policy explicitly stating that blocks from arb motions shouldn't be unblocked but WP:NEVERUNBLOCK only states that remedies should never be unblocked? Terasail[✉️] 00:15, 7 January 2023 (UTC)[reply]
    @Terasail that's part of the larger issue - the admin and blocking policies are not really the same here - and really they shouldn't be redundant. I think ideally the admin policy should just refer that special situation blocks exist and have special rules, and everything about "blocking" belongs in the blocking policy. I think barkeep49 is trying to make an incremental positive improvement that will be easier to pass than to have a larger more encompassing discussion right now. — xaosflux Talk 00:22, 7 January 2023 (UTC)[reply]
    I have no concerns with more copy editing of the other bulletpoints and am supportive of even more alignment of the language between ADMIN and BLOCK - ideally through transclusion so that they could never become out of sync again. But unlike the other issues we're talking about - which is about what not to do - the bulletpoint in ADMINS gave affirmative permission to do something. If some admin were to reverse an unblock of Arbcoms and go "But it wasn't in ADMIN" they wouldn't get very far. But I think it has been entirely possible for good faith admin to rely on the language in ADMIN to do something that isn't ultimately endorsed by the community any more. Fixing that is my priority. Best, Barkeep49 (talk) 01:32, 7 January 2023 (UTC)[reply]
  • Slight tweak to Isaacl improvement, keeps it succinct, preserves one of the ideas above that such blocks can't be made discretionarily:
    • Blocks made by the Arbitration Committee: Blocks authorized by the Arbitration Committee must have a summary containing "For the Arbitration Committee", "Appeal is only to the Arbitration Committee", or "{{ArbComBlock}}". Administrators must only place, change, or remove such blocks with the prior, written consent of the committee. (See also: Wikipedia:Arbitration/Policy#Appeal of decisions.)
    • xaosflux Talk 18:48, 7 January 2023 (UTC)[reply]
      I like the brevity but using the word "change" introduces a complication we've seen with changing CU blocks. Admins can 'tighten' blocks, often for good reason and sometimes in a hurry, by yanking email or TPA - typically for someone abusing email or their talk page. Blocks may also be extended, even though it may not always be particularly recommended. With CU blocks, if the latest entry is not made by a checkuser, then you have to trawl through the block log to understand its origin. You can end up with layers of blocks. The CU and OS policies are not perfect and are inconsistent in this regard, but do ultimately veer towards "loosen" rather than "change" or "alter". I'd suggest retaining this approach. -- zzuuzz (talk) 19:40, 7 January 2023 (UTC)[reply]
      @Zzuuzz how about place, reduce, or remove? This wouldn't apply to "CU" blocks (CU and OS are in their own bullets that aren't being proposed for change today), only "ARBCOM" blocks. However there is the situation where arbcom blocks someone, then normal community response is needed to further block them. — xaosflux Talk 20:13, 7 January 2023 (UTC)[reply]
      "Reduce" works for me. I was merely highlighting why CU blocks already avoid the word "change". They and Arbcom blocks are very similar in nature. -- zzuuzz (talk) 20:32, 7 January 2023 (UTC)[reply]
      "Place, reduce, or remove" sounds good to me, too. It might be a generational thing: each time I read "the prior, written consent" I want to follow it up with "of Major League Baseball".) isaacl (talk) 22:59, 7 January 2023 (UTC)[reply]
    I think ArbCom has a fair amount of discretion, under ARBPOL, for how it words its blocks and so the must language (which I hadn't noticed before) sets-up a potential conflict between ADMIN and ARBPOL that the previous language did not. Best, Barkeep49 (talk) 23:27, 7 January 2023 (UTC)[reply]
    @Barkeep49 agree, I would say they "must" identify that they are arbcom blocks, but shouldn't require exact verbiage - can tweak that. — xaosflux Talk 00:02, 8 January 2023 (UTC)[reply]
    @Xaosflux I agree they must clearly state they are blocks by the Arbitration Committee in some form or fashion but that particular method needn't be specified in the policy. Best, Barkeep49 (talk) 00:06, 8 January 2023 (UTC)[reply]
    Sure, it could be something like "Blocks authorized by the Arbitration Committee must include a clear indication of their source, such as..." isaacl (talk) 01:33, 8 January 2023 (UTC)[reply]
The discussion above is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.

Allow registered editors to use vpn (open proxies)

Currently, WP:PROXY states about itself, "This policy is known to cause difficulty for some editors who may need to use open proxies". I have experienced said difficulty and whenever I try to use my vpn I get the Wikipedia notice that I am not able to edit because of it. The rationale of the policy states, " open proxies are often used abusively. MediaWiki, the wiki software that powers Wikipedia, depends on IP addresses for administrator intervention against abuse, especially by unregistered users." Why not let registered editors use vpn (open proxies)? When I use an online website with interaction with other users, oftentimes I can block any given user, I don't need their ip. I don't see why Wikipedia cannot do the same by just blocking the account without resorting to the ip.

The current policy and technical actions of blocking the use of open proxies by registered users seems to be unreasonable. Placing bureacratic hurdles to be able to use one seems to be also unneeded and unreasonable when it comes to said registered editors.

Many people who want to contribute to Wikipedia are probably just enthusiastic about editing here but they may not have much idea in what kind of serious problems that can even destroy their life they can get in by the simple act of editing.[1] [2]

As a balance between vandalism by anonymous users and the safety of editors, Wikipedia should allow unrestricted vpn (open proxies) use by registered editors. Thinker78 (talk) 17:11, 9 January 2023 (UTC)[reply]

@Thinker78: This is pretty much a nonstarter. The IP is needed to help prevent block evasion via WP:SOCKPUPPETRY. If you really need edit behind a proxy, then simply just request WP:IPBE.
It's also not true that other websites don't block VPNs. Netflix routinely does this as well (albeit for very different reasons). –MJLTalk 19:52, 9 January 2023 (UTC)[reply]
I never said that "other websites don't block VPNs". Besides, what's the percentage of registered editors who have been blocked? What's the percentage of those blocked editors who could cause really harm to the project for suckpuppetring as opposed to any ip user or new account? Also, for any issues with a new sockpuppet account, pages can be protected.
According to WP:IPBE there are only 806 editors with the block exemption out of the millions of editors in Wikipedia. That's not a very successful statistic of the program. The balance to be made is between sockpuppetry and the safety of editors. Thinker78 (talk) 22:09, 9 January 2023 (UTC)[reply]
The people who are blocked by this are disproportionately from developing countries. See m:Talk:No open proxies/Unfair blocking for some examples.
This is going to become a bigger problem. Blocking everyone who uses Apple's iCloud Private Relay is going to cut into the English Wikipedia's core editor base. We're asking people to choose between disabling privacy features on all websites, or not being able to edit. Google Chrome, which is the most popular web browser among editors, is likely to ship something similar in the next year or two. MediaWiki (the software that we use) may have to stop focusing on IP addresses and move to another system, like a Device fingerprint.
I know that there's been talk among the devs and product managers about this problem recently (also, off and on for at least ten years). The one thing that any registered editor could do to help in the short term is to turn on the "IP Info" item in Special:Preferences#mw-prefsection-betafeatures. I've found this answers most common questions (e.g., what part of the world is this IP editing from?), and it's really handy on the history page. Please try it out, and provide your feedback to the team, so they can get this initial project wrapped up. Whatamidoing (WMF) (talk) 03:38, 10 January 2023 (UTC)[reply]
I live in Guatemala and many times when I try to edit from my cell phone using data and not wifi I find that I am blocked. For some reason the ip address assigned to my phone by my ISP is not of the liking of the blocking code or the dev who placed the restrictions. Basically people are blocked from editing from cell phones only because they live in Guatemala. Thinker78 (talk) 15:49, 10 January 2023 (UTC)[reply]
We have the same problem in the US. If you use T-Mobile (the second biggest mobile phone service in the US), then you can't edit from your phone. Admins have blocked all the IPs. Whatamidoing (WMF) (talk) 22:42, 11 January 2023 (UTC)[reply]
With respect it's nowhere near the same problem. T-Mobile is just about never hard blocked. You just need to log in (and if you think that needs softening then you'd be wrong). -- zzuuzz (talk) 22:51, 11 January 2023 (UTC)[reply]
That sounds like something they should take up with their VPN client provider. Many clients allow you to whitelist destinations. — xaosflux Talk 15:56, 10 January 2023 (UTC)[reply]
That might be possible if we could safely assume that they have control of the VPN (and, e.g., not the VPN that 30% of internet users say they are required to use for work), and if we assume that someone who wants to add a sentence or correct an error in an article has enough skill to know how to do that. Just from those two groups, I'd guess that less than half of people are able to do this in practice. Whatamidoing (WMF) (talk) 22:37, 11 January 2023 (UTC)[reply]
Whereas other sites require some form of identity verification, whether through email, phone, payment, social media or other checks and metrics, Wikipedia requires none of that. The requirements for an account here are basically non-existent. For this reason, users who are merely registered or unregistered are usually more similar than you might think. -- zzuuzz (talk) 20:17, 9 January 2023 (UTC)[reply]
This three-part series on proxy blocks from Vermont on Diff may be relevant: first part, second part, third part. 🐶 EpicPupper (he/him | talk) 05:25, 10 January 2023 (UTC)[reply]
You are linking to the same article for the second and third parts. 2603:8000:FC00:800:1055:C75E:A3A:3DE (talk) 17:39, 10 January 2023 (UTC)[reply]
Good catch, thanks. 🐶 EpicPupper (he/him | talk) 01:51, 11 January 2023 (UTC)[reply]
Those are great blog posts. BTW, the blog's open to volunteers. If you know something about this problem (or another subject) and want to write it up, please contact User:RAdimer-WMF. Whatamidoing (WMF) (talk) 22:44, 11 January 2023 (UTC)[reply]
I have been caught out on this a few times myself due to the iCloud relay before giving up on mobile or iPad editing. I was initially thinking that the idea was good but once reading the replies made me realise the issue was more complex. My immediate thought was that, perhaps, it could be Extended Confirmed required to use an open proxy (or carrier grade NAT or similar). If an extended confirmed user gets blocked then their user created for block evasion would still need to reach EC before being able to bypass anything. Gusfriend (talk) 02:09, 11 January 2023 (UTC)[reply]
Unfortunately we already have disruptive editors who game extended confirmed so that they can edit disruptively on articles that are EC protected. Making it so that upon granting of EC the editor is also given the equivalent of IPBE would make it harder for CheckUsers to identify repeat offenders and block any sleepers they may have pre-prepared. Sideswipe9th (talk) 01:32, 15 January 2023 (UTC)[reply]
How do they game extended confirmed? Besides, the issue cannot just be thought about restricting editors but there has to be a meaningful balance with the safety and privacy of editors online. Thinker78 (talk) 03:22, 15 January 2023 (UTC)[reply]
Same way disruptive editors game autoconfirmed. I'll not say more than that per WP:BEANS, just that the difference between the two types of gaming is one of volume and time. If you edit any of the more controversial topic areas though, I'm fairly positive you will have seen this behaviour.
I agree that there is a balance to be struck between safety and privacy, but I do not agree that granting all extended confirmed users meets that balance. IPBE is closer to balancing those needs, though it is also not without its problems. Sideswipe9th (talk) 03:31, 15 January 2023 (UTC)[reply]
"it could be Extended Confirmed required to use an open proxy (or carrier grade NAT or similar)" – I agree strongly with this. It's time Wikimedia left the 2000s and entered the 2020s on security, but we don't want randos being able to hide their IPs from us.  — SMcCandlish ¢ 😼  23:46, 13 January 2023 (UTC)[reply]
We already have WP:IPBE though. If you are effected, then apply. If lots of people get denied this, then we need to rethink, but if we're just complaining about it without applying it's a bit of a non-starter. Lee Vilenski (talkcontribs) 01:10, 15 January 2023 (UTC)[reply]
As I mentioned previously, "According to WP:IPBE there are only 806 editors with the block exemption out of the millions of editors in Wikipedia." Thinker78 (talk) 03:27, 15 January 2023 (UTC)[reply]
That's not a relevant statistic though as the vast majority of those editors have no need for IBPE (or had no need at the time they edited). We have a current system (IPBE) and so before replacing it with something else we need to determine whether it is fit for purpose and, if it isn't, whether it needs modifying or replacing. Some questions that should help determine whether it is fit for purpose are:
  1. How many people have a need for IBPE (or a replacement)?
  2. Of those, what proportion are good faith editors? (because at least some bad faith editors would definitely benefit in ways we do not want)
  3. How many of those good-faith editors know to ask for IBPE?
  4. Of those good-faith editors that ask, how many are approved?
I don't pay a lot of attention to the IPBE discussions that happen on the Functionaries mailing list, but from what I remember the answer to 4 is that it is at least a majority and probably a significant one. Thryduulf (talk) 12:45, 15 January 2023 (UTC)[reply]
That statistic doesn't include admin, who have the right by default. The question is how many people are being denied access via a IPBE request? Lee Vilenski (talkcontribs) 13:47, 15 January 2023 (UTC)[reply]
@Lee Vilenski admnins have "half" of IPBE; they can bypass blocks but can't use tor. — xaosflux Talk 17:03, 15 January 2023 (UTC)[reply]
What's your standard to weigh whether someone needs IBPE? Thinker78 (talk) 17:47, 15 January 2023 (UTC)[reply]
My personal standard is that IPBE is needed only when someone is unable to contribute (safely) without it and will use it only to contribute in good faith. The second is obviously hard to judge objectively, especially for new users, but for existing contributors your contribution history is usually a fairly reliable guide. (Note this is not an official position of the functionaries or any other group). Thryduulf (talk) 15:36, 16 January 2023 (UTC)[reply]
How is the determination if someone is unable to contribute safely? Examples? Thinker78 (talk) 18:44, 16 January 2023 (UTC)[reply]
There is and can be no single criterion for this as it depends a lot on individual circumstances, but an editor in a first world country wanting to write about first world topics can do so safely without IPBE in pretty much every case (but exceptions may exist). In contrast an editor in Iran wanting to write about contemporary Iranian politics will plausibly need to use a proxy for their own safety. Thryduulf (talk) 20:46, 16 January 2023 (UTC)[reply]
I think that standard is unreasonable. For starters, security experts recommend using vpn when connecting to a network one doesn't trust (it could be public wifi, at home with relatives or roommates one doesn't trust, or a snooping ISP who wants to sell data).[3]
In addition, regarding first world countries, you probably didn't hear of the time when the government of France forced an editor of the French-language Wikipedia, and president of Wikimedia France, under threat of detention and arrest, into deleting an article about Pierre-sur-Haute military radio station.
This points out that even in first world countries, specially in the US, people are subject to violations of their human and legal rights regularly by authorities. Police regularly arrest people for legally protected things, like taking pictures in public places, for having a bumper sticker in the truck.[4] or for criticizing public officials[5]
Finally, some US law enforcement authorities are deeply corrupt and can even murder people for criminal organizations.[6] Thinker78 (talk) 23:03, 16 January 2023 (UTC)[reply]
[4]. 🐶 EpicPupper (he/him | talk) 23:46, 16 January 2023 (UTC)[reply]
@EpicPupper that's like the dilemma as to whether to have bodyguards or not. How would one know the bodyguards are not bought out or tracking one's every move for nefarious purposes or even to plan to takeover one's businesses? Some bodyguards are really great people, but others probably not so much.
The issue then becomes of choice. The normal advice is to use an open proxy if one thinks the connection is not safe. Some providers of vpn may be dishonest but others are not. Whether its use gets to be beneficial or not really depends on individual circumstances. Thinker78 (talk) 19:57, 17 January 2023 (UTC)[reply]
I'll add my own personal criteria, which is slightly different to Thryduulf, and probably on the more liberal side of written policy. The user generally needs to be a) unreasonably affected by blocks, and b) not a sockpuppet. There's obviously some judgement which goes into that. I'm not hugely sympathetic to the first world problem of not being able to edit through Apple relays from your latest iPhone due to some false sense of security. Be reasonable: Just turn it off. It's not going to get you arrested, and it will help checkusers. Also, trust me, you don't want to be using the same IP addresses and user agent as some banned troll making death threats. I'm open to persuasion in individual cases though. For users in countries strongly affected by P2P blocks, or in countries with serious security issues, I'll often grant IPBE without even a request, subject to my second criterion. As for who is not a sockpuppet? it's probably one of those things you know it when you see it. I've seen many many extended-confirmed socks, some of whom request (and sometimes get) IPBE. There's no real threshold, short of several years and many thousands of edits, where I'd place any automatic grant. For one thing, we have a plague of accounts getting compromised, and having IP blocks can help when that happens. I find it interesting looking through all the IPBE grants made by all the admins above. I count maybe two in the last several years. There's probably several factors at play there. Lack of requests is probably a leading factor. Also, talking of statistics, the number quoted is just a snapshot. Many IPBE grants come and go as, like a lot users, they're usually temporary. -- zzuuzz (talk) 20:27, 16 January 2023 (UTC)[reply]
Zzuuzz, as per my previous reply, there are no guarantees that no one gets arrested or worse for what they contribute in Wikipedia in the US. Thinker78 (talk) 23:07, 16 January 2023 (UTC)[reply]
If the authorities are watching what you're doing through your account and HTTPS then you have bigger problems than your IP address. I'll wager that using Apple services won't help your situation either. Most of what you mention is unrelated to IP addresses. I'll just add that I have granted IPBE for these types of concerns. They're not common, and are rarely plausible. -- zzuuzz (talk) 23:39, 16 January 2023 (UTC)[reply]
It's more about people not in the radar of despotic or corrupt authorities not getting in their radar. And if they are in their radar, make it more difficult to get snooped on and not completely forget about cybersecurity. I know that Wikipedia has a secure connection but it has happened to me that the network transaction to connect to it sometimes is reset in the middle of it and the connection becomes unsecured for a couple of seconds, enough to reveal the contents of whatever I am doing at the time. Thinker78 (talk) 20:03, 17 January 2023 (UTC)[reply]
  • I don't this this is a great idea as-is, but perhaps something. We could make some intermediary access, that only has (ipblock-exempt), and does not have (torunblocked, sfsblock-bypass) - but would need to decide on some bar for it... so it would need to be what does the community and admin who have to deal with this is a low enough bar? I don't think econf is high enough - maybe something like 1000edits and 180 days with autopromoteonce. If the bar is so high that it would be useless, then it is useless to bother though. — xaosflux Talk 17:09, 15 January 2023 (UTC)[reply]
    Maybe 90 days? Half a year seems like quite a high bar. 90 days would already be 3x XC and more than inconvenient for anyone trying to game the system. Terasail[✉️] 17:33, 15 January 2023 (UTC)[reply]
    According to Wikipedia:List of Wikipedians by number of edits, only 0.25% of editors are extended confirmed. That's 1 in 400. Now the questions are, what range of number of edits do you think is entry level bar for ipblock exempt? What's the global share of edits performed by said segment of editors? To have a rough measure of good faith edits, what's the percentage of said editors who have been blocked indefinitely (keep in mind that even User:Koavf was at one point blocked indefinitely)? Thinker78 (talk) 17:44, 15 January 2023 (UTC)[reply]


Bare references

The following is one of the rules for the article creation process: "Articles should reflect only what reliable sources have said about the topic, and all articles need references to reliable sources." However, some of the draft articles are accepted although they have bare references. So, how do the reviewers implement this rule without knowing the reliability of the references when these are given in a bare form? I suggest that those articles created through this process should have full references. It would make it possible to implement this rule and also, to reduce the number of the articles with bare references. Filling bare references is a very hard and time-consuming activity for the editors. Best, Egeymi (talk) 13:05, 7 January 2023 (UTC)[reply]

A bare reference to a reliable source is still a reference to a reliable source and therefore does the job. Any reviewer who can't tell if a bare reference is to a reliable source or not frankly shouldn't be reviewing articles at all, so I don't see any problems here. IffyChat -- 15:14, 7 January 2023 (UTC)[reply]
One could place a tag like {{Cleanup bare URLs}}, but it's a yellow tag, which isn't deemed critical, though should eventually be addressed.—Bagumba (talk) 15:44, 7 January 2023 (UTC)[reply]
Bare refs are references. The citation style is more to do with WP:LINKROT than not being suitable references. WP:REFILL is a thing as well that'll sort out a lot of these. Lee Vilenski (talkcontribs) 16:47, 7 January 2023 (UTC)[reply]
@Egeymi, if you don't like running REFILL, then try out the visual editor. If you find a bare URL in ref tags, then it offers a "convert" button. It can fill in most citation templates automatically, though I don't think either system does a great job with PDFs. Overall, filling bare references isn't difficult or time consuming any longer. Whatamidoing (WMF) (talk) 03:41, 10 January 2023 (UTC)[reply]
@Whatamidoing (WMF) Thank you for pinging me, Egeymi (talk) 06:07, 10 January 2023 (UTC)[reply]
Per all of the above, draft reviewers should not be holding up drafts that would pass WP:AFD and be kept. Citation format is not a valid reason to delete an article, ergo, it's not a reason to deny passing a draft into the article space. Bare references are at once sufficient to prove a subject is notable, and also in need of additional work. It's good enough for the mainspace, though as noted, someone should feel free to fix them!--Jayron32 15:53, 9 January 2023 (UTC)[reply]
Thank you for all views, I just wanted to know if these bare urls may be given as full references at the beginning of the process by the editor who submits the article. I didn't say anything about the deletion of such articles with bare urls. Anyway, thanks. Egeymi (talk) 06:07, 10 January 2023 (UTC)[reply]
@Egeymi Sometimes Citation Bot can also fill in bare references, but I am not sure if it can be used on one off articles. Rlink2 (talk) 22:04, 10 January 2023 (UTC)[reply]
@ Rlink2 Thank you, I use all available tools. Egeymi (talk) 05:34, 11 January 2023 (UTC)[reply]
Refill is actually one of the suggested tools for reviewers at AfC and next to the CopyVio tools at the top of the article. If there is more than a couple of bare refs (most of the time if there are any) or uncombined references then I run it to fix the issue. Gusfriend (talk) 02:12, 11 January 2023 (UTC)[reply]
Thanks, Egeymi (talk) 05:34, 11 January 2023 (UTC)[reply]
It is not too hard to click on a bare reference to see if it immediately is a problem or not - for example, this needs to be done with any forbes.com links as staff and contributor links have the same appearance. But we should definitely not pushing draft writers for not fleshing out a reference format as long as the link is there. Masem (t) 05:03, 11 January 2023 (UTC)[reply]

Removal of vague and relative wording from weather/climate sections on places

From Portland, Maine § Climate:

Portland has a humid continental climate (Köppen: Dfb, closely bordering on Dfa), with cold, snowy, and often prolonged winters, and warm, relatively short summers.

From Livermore, California § Climate:

Livermore features hot, dry summers and mild to cool winters with occasional rainfall (Köppen climate classification Csa).

I propose removing the words "cold", "warm", "dry", etc. because they are entirely relative to the climates of other locations, and since the articles already have Template:Weatherbox, these words aren't giving any new information. Numberguy6 (talk) 21:46, 7 January 2023 (UTC)[reply]

Meh… These are words that are commonly used when discussing climate zones. How would you suggest we convey the same information without using them? Blueboar (talk) 22:49, 7 January 2023 (UTC)[reply]
Although clearly what's considered cold in Norway is different than what's cold in Bermuda, I think readers are able to contextualize the climate description to the general region in question, and thus it's a helpful shorthand. isaacl (talk) 23:02, 7 January 2023 (UTC)[reply]
  • It's basically a quote/paraphrase of the Köppen climate classification itself; which is useful because most people don't know what Dfb means, but they can get a sense for what "Cold, snowy, prolonged winters" means. --Jayron32 15:50, 9 January 2023 (UTC)[reply]
  • I don't mind the descriptors given above. I have problems using things like "bitterly" and "extreme". CambridgeBayWeather, Uqaqtuq (talk), Huliva 18:08, 10 January 2023 (UTC)[reply]
  • Those descriptors are far more useful than "Dfb, closely bordering on Dfa". CMD (talk) 05:10, 11 January 2023 (UTC)[reply]
  • I don't see how this is workable. Aside from the arguments already made above for retaining the option for such language as simple terms (be they admittedly relative and imprecise) for the benefit of the average reader, there's also the fact that these descriptors are often going to align with the language employed by the relevant reliable sources for a location's climate/average temperatures, so I think that in and of itself creates an unsurmountable obstacle for this proposal. That said, I don't see a problem with editors attempting to make better and more nuanced descriptions where feasible SnowRise let's rap 18:42, 13 January 2023 (UTC)[reply]
  • This isn't something that lends itself to codification in policy. It depends on what the sources say, so, if there is a dispute about any particular article, it should be decided by consensus on the article talk page. Phil Bridger (talk) 19:31, 13 January 2023 (UTC)[reply]
How is this helpful to anyone? Are terms as simple and objective as “hot” and “cold” now going to be (wrongly) considered “non neutral” in the way other simple, objective terms like “far right” and “pseudoscience” are? We aren’t describing climates as “awful” or “nice”; some parts of the Earth are just hotter than others. Dronebogus (talk) 11:31, 16 January 2023 (UTC)[reply]
I have worked in Livermore, California many times over a 30 year period, and the description above is accurate and neutrally written. I support keeping this type of prose in city articles. Cullen328 (talk) 18:16, 16 January 2023 (UTC)[reply]

Proposals at WT:article size

 – Pointer to relevant discussion elsewhere.

I'm in the process making a couple of proposals to change WP:Article size. See the first substantial draft proposal Wikipedia_talk:Article_size#Maintainability_issues, and watchlist the guideline for more proposals to come if you're interested. —Femke 🐦 (talk) 10:06, 14 January 2023 (UTC)[reply]

 – Pointer to relevant discussion elsewhere.

WP:GA has started the new year with a discussion on covering a range of proposals to change policy and update guidelines. The discussion has been live for the last 2 weeks with approximately 20 different proposals currently being discussed. If your interested in participating or offering your own proposals then follow the link to Wikipedia talk:Good Article proposal drive 2023. 🏵️Etrius ( Us) 03:02, 15 January 2023 (UTC)[reply]

Should non-free images be allowed in search results?

 You are invited to join the discussion at Wikipedia talk:Non-free content § Non-free images in search results (redux). {{u|Sdkb}}talk 19:46, 16 January 2023 (UTC)[reply]