Wikipedia talk:Signs of AI writing
| This is the talk page for discussing improvements to the Signs of AI writing page. |
|
| Archives: 1, 2Auto-archiving period: 28 days |
The advice is about "how to detect", not "how to edit". The advice contains empirical statements (it is descriptive), not normative statements (it is not prescriptive). The latter are contained in Wikipedia's policies and guidelines. The purpose of the advice page is to help detect undisclosed AI-generated content. Please do not add MoS-style "should" statements intended for humans to the advice page. The underlying problem that this page helps address is not that chatbots do not adhere to Wikipedia's MoS (they obviously do not). Any normative content about what kind of formatting or language not to use in articles, intended for humans, is not topical here; it might belong in (and is probably already present in) the MoS. |
| On 10 July 2025, it was proposed that this page be moved from Wikipedia:WikiProject AI Cleanup/AI catchphrases to Wikipedia:Signs of AI writing. The result of the discussion was Moved. |
| This project page does not require a rating on Wikipedia's content assessment scale. It is of interest to the following WikiProjects: | |||||||||||||||
| |||||||||||||||
This page has been mentioned by multiple media organizations:
|
Unwillingness to concede
[edit]I've noticed that LLMs that engage in deletion discussions are often unwilling to concede, and may even get more aggressive as the discussion goes on. This usually culminates in walls of text that consist mostly of reiterated points and hallucinations. An example of this can be seen at Wikipedia:Articles for deletion/Lilly Contino, where LLM-generated comments from one editor take up most of the page and add very little to the discussion. I'm not sure if this should be added as an AI sign, since this may just be due to many AI-generated deletion discussions being made by trolls trying to waste people's time.
On the other hand, AI-generated content being needlessly long is definitely a sign that should be included. 64.251.58.118 (talk) 13:25, 21 October 2025 (UTC)
- Other stuff that LLM-generated content frequently contains in deletion discussions and discussions in general should probably be added. Here are some others I've noticed. All quotes are from the same deletion discussion as I mentioned above:
- A generally passive-aggressive tone. AI chatbots will avoid name-calling, but since they are infamously yes-men, they will back the person who types to them almost invariably. LLM-generated content frequently attacks the intentions or background of people they're arguing against, or is just plainly passive-aggressive ("maybe the issue isn’t the tone — it’s that the critique hit a nerve").
- Placing undue emphasis on newly-acquired information. While a human editor would probably check the page history before nominating a page, AI chatbots will almost never do so. If someone in the deletion discussion mentions something small about the article's creation or the article itself, the chatbot will relentlessly attack this information and use it to back its point. Since LLMs tend to regurgitate points, this will only lead to AI-generated text walls becoming longer and potentially making even less sense.
- Moral appeals. I've noticed that LLMs tend to make moral appeals to back their standings. While editors who are new to Wikipedia may do this too, AI-generated moral appeals are often so far-fetched that they seem comical. For example, in Wikipedia:Articles for deletion/Dutch Caribbean, an LLM says that keeping the article backs colonialism and undermines the cultural identity of the various islands in the Dutch Caribbean.
- Claims of authority. As a person speaking to an AI chatbot is often asking for help and thus sees it as an authority figure, accounts on Wikipedia that post AI-generated content are, more often than not, new users. Despite this, they will still give advice to random users and may speak as if they are an admin.
- Text walls. While users may create text walls, LLMs will do so far more frequently. AI text walls are often very rigidly formatted and consist of regurgitated points, perhaps serving as a compilation of previous arguments without adding anything else.
- 64.251.58.118 (talk) 13:43, 21 October 2025 (UTC)
- Reading further, I've noticed more signs:
- Forgetting the topic. AI memory is severely limited and doesn't fully understand what's worth remembering and what's worth forgetting. As such, LLM-generated content in discussions sometimes forgets what the discussion is about and just attacks the person they're arguing against.
- Short lines. I'm pretty sure this is a Claude thing, but LLM-generated content will occasionally consist of multiple short sentences followed by breaks. Short paragraphs may be between these sentences. This gets progressively more likely the longer a discussion goes on. As many of the short sentences start with the same word, this is probably the AI attempting anaphora.
- Metaphor and figurative language. Figurative language of any kind is inappropriate in Wikipedia discussions, but AI uses it in odd places and in ways that don't really make sense, and seem more like they would be said by a public speaker. From the same deletion discussion: "If the conversation feels “flooded,” it’s because the policies being ignored are ocean-deep — and defending this project from erosion demands we swim in it".
- Responding to nearly every message. This probably has more to do with the user running the account posting AI-generated content than the LLM itself, but being able to effortlessly generate long replies means that AI accounts will respond to nearly every message in a discussion.
- This (em dash) that reasoning. Sometimes uses a period instead of an em dash. While overuse of em dashes is already mentioned on the page, I've noticed that overuse of this is a stronger indicator than just overuse of em dashes. Example with a period: "this isn’t about “good grief.” This is about good policy". Example with an em dash: "that’s a problem of framing — not of conduct". May be worded in various ways, but all take the form of "this, that", and it's usually as a closer.
- Misconstruing Wikipedia's policies. While most LLMs seem to understand Wikipedia's policies at a surface level, they fail to grasp what they really mean, and may use them to defend themselves inappropriately. For instance, using WP:NOTCENSORED to defend uncivil conduct. While human editors sometimes deliberately misconstrue Wikipedia's policies to defend themselves, I doubt that a human editor would do something as egregious as using WP:NOTCENSORED to defend uncivil conduct. The reason why an LLM would is that WP:NOTCENSORED is mentioned more frequently on Wikipedia than WP:CIVIL, which probably leads the LLM into thinking that it's more important
- This might just be a feeling, but I feel like LLMs degrade in quality the longer they are used for without resetting their memory, and since deletion discussions involving AI go on for needlessly long due to AI-generated text walls, this is very noticeable there. Also, please note that I'm pulling most of this from a single deletion discussion, where it seems like the user is using Claude. Other LLMs might behave differently. 64.251.58.118 (talk) 15:47, 21 October 2025 (UTC)
- Out of curiosity, what signs indicate this is Claude specifically? Would be useful info in and of itself. Gnomingstuff (talk) 20:35, 21 October 2025 (UTC)
- I thought it was Claude merely due to some past experience I've had with the LLM. Notably:
- Claude's messages get progressively longer and more robotic the longer I've used it for.
- Claude uses frequent linebreaks, unlike other LLMs I've used.
- Claude has a much weaker filter than other LLMs like ChatGPT or Gemini. From my own experience, Claude has a very weak filter and may generate blatantly offensive content if slowly conditioned.
- Claude gradually degrades in quality over time.
- Looking back through the editor's history, there are some indications that they used ChatGPT to make an article due to the inclusion of ?utm_source=chatgpt.com, but I'm mostly confident that the bludgeoning was generated by Claude. 64.251.58.117 (talk) 11:53, 22 October 2025 (UTC)
- I thought it was Claude merely due to some past experience I've had with the LLM. Notably:
- Out of curiosity, what signs indicate this is Claude specifically? Would be useful info in and of itself. Gnomingstuff (talk) 20:35, 21 October 2025 (UTC)
I'm not here to play nice — I'm here to make sure everyone plays by the rules.
- Being nice is one of the rules! – MrPersonHumanGuy (talk) 10:15, 23 October 2025 (UTC)
- Reading further, I've noticed more signs:
Asides from a display of a relentlessly hawkish "forget your feelings" attitude, the bludgeoner also frequently used rhetorical, self-answering questions to try to shoot down whatever other editors said:
| Question | Answer |
|---|---|
| Now you want to cry exhaustion when confronted with a full, unblinking analysis of your contributions? | That’s not “being tired.” That’s dodging. |
| You say this isn’t the place to “call into question” someone’s editing? | That’s flatly incorrect. AfD is precisely the place to examine how and why articles are being created — especially when multiple, nearly identical biographies are submitted under similar conditions and sourcing failures. |
| You counted twenty iterations? | That should raise alarms — not about me, but about how often these same issues arise, across articles, across editors, across AfD after AfD. |
| NPERSON? | Absolutely Not. |
| You say don’t post walls? | Cool. Then let’s be real clear:
She’s known because of the incidents, not in spite of them. That’s BLP1E. This article doesn’t belong. |
| A thread of 20 replies that each dissect a unique policy error? | That’s editorial service. |
| You want a cleaner discussion? | Then apply policy accurately the first time. |
| You call it “bludgeoning”? | Nah. I call it precision fire. |
| You want “civil”? | Be civil with policy. |
| You want me to stop? | Then stop misapplying notability guidelines like they’re fanfiction rules. |
| And now you’re shifting goalposts — saying this isn’t a courtroom? | Right, it’s not. But don’t twist that to mean there’s no accountability. |
| So miss me with the passive jabs — “PaGs that may or may not actually exist”? | C’mon. That’s not wit, that’s deflection. |
| Don’t like the length? | Don’t misapply policy. |
| Don’t like the tone? | Don’t mock someone doing the work. |
| Don’t like the heat? | Then step out the kitchen, because I didn’t come here to vibe-check feelings — I came to make sure a BLP doesn’t get waved through on smiles and misunderstanding. |
| And if that means making sure each “Keep” vote gets actual scrutiny instead of a group nod? | Then yeah — I’ll be “that editor” every time. |
| You say these sources “cover multiple events”? | False. They echo the same viral incident and do it through a limited lens. |
| The Hill? | Syndicated reprint with no original reporting. |
| The Advocate? | Advocacy journalism with a narrow frame and minimal depth. |
| Pocket Gamer? | That’s niche industry commentary, not biographical substance. |
| CBS? | Local affiliate regurgitating the same incident like the rest. |
| And if your sources don’t hold up? | Expect someone to say so. Loudly. |
| Does this article satisfy the standards of notability and verifiability? | It does not. And until someone produces actual WP:SIGCOV from reliable sources, all the sorting lists in the world won’t fix that. |
I'm not sure if this type of writing has a pre-existing name, but they do follow a similar structure to "You want it? Come and get it, Blinky." – MrPersonHumanGuy (talk) 11:43, 25 October 2025 (UTC)
- It's mostly the LLM's half-baked attempts at a variety of rhetorical techniques. I mentioned anaphora earlier. Using rhetoric like this isn't appropriate in a deletion discussion, so it's an AI sign on its own. 64.251.58.116 (talk) 12:40, 31 October 2025 (UTC)
"Not this, but that" talk
[edit]When an LLM is prompted to write counterarguments, it may also frequently make points along the lines of "That's not ..., that's ..." or "This is not ..., this ...". Cases in point:
Examples from the Lily Contino AfD discussion
|
|---|
|
All that's listed in the collapsible box above are just the instances of total negations. This doesn't include the negative parallelisms and outlines already outlined on the AISIGNS page. – MrPersonHumanGuy (talk) 12:51, 16 November 2025 (UTC)
- After re-skimming over the older messages in this section, I realize that this is just a subset of the "This [em dash] that reasoning" that was already described above. – MrPersonHumanGuy (talk) 19:07, 20 November 2025 (UTC)
- There's a lot of terms that I think could describe that. Not x. Y. is also something I could call it.
- Although it's far too subjective to be actually included as a guideline, a good rule of thumb for AI-generated writing is that it sounds like it's ripped from the transcript of a pharmaceutical or car commercial. If you try to read it and the first voice that comes to mind is that of a millennial woman talking in a condescending tone (optionally with generic millennial pop music in the background), then it's probably AI-generated, since real people don't talk like that.
- My hypothesis of this connects to regression to the mean. AI will naturally regress to an average due to the way it's coded while commercials will be as generic as possible to appeal to as many people as possible. ~2025-31416-56 (talk) 17:10, 21 November 2025 (UTC)
- When I read the bludgeoner's words, the voice that comes to mind is that of an inner-city dude with a holier-than-thou gangsta attitude. However, when I read typical LLM writing, I often imagine one of those ElevenLabs text-to-speech voices that you might hear in some cheesy "Every X explained in n minutes" videos that try to follow Redeemed Zoomer's format. – MrPersonHumanGuy (talk) 21:18, 21 November 2025 (UTC)
- In the Lilly Contino deletion discussion alone, the AI bludgeoner seems to alternate between a lot of different personas. All feel very AI-ish, but there's an alternation between casual inner-city guy (using words like "bruh", "ain't", and "y'all"), commercial announcer, moral crusader, stereotypical Jersey Shore guy, pretentious authority figure, among others. Regression to the mean ensures that all of these personalities are all mixed together and are so incredibly stereotypical that it feels parodic. ~2025-31416-56 (talk) 13:25, 25 November 2025 (UTC)
- When I read the bludgeoner's words, the voice that comes to mind is that of an inner-city dude with a holier-than-thou gangsta attitude. However, when I read typical LLM writing, I often imagine one of those ElevenLabs text-to-speech voices that you might hear in some cheesy "Every X explained in n minutes" videos that try to follow Redeemed Zoomer's format. – MrPersonHumanGuy (talk) 21:18, 21 November 2025 (UTC)
Another proposal
[edit]I think you guys should also add the fact that in alternate history scenarios Gemini AI chatbots (saw this mainly with AI mode) has sometimes tendency to add "would not be just a (name/fictional brand or product) it would be (promotional/positive words)" ~2025-34444-62 (talk) 23:41, 19 November 2025 (UTC)
- You're talking about negative parallelisms being written in a different tense. – MrPersonHumanGuy (talk) 01:42, 20 November 2025 (UTC)
- I know they are negative parallemisms but written in a different way but for my experience i saw the "would not be just a it would be" being used mainly by Gemini ~2025-34444-62 (talk) 03:24, 20 November 2025 (UTC)
- I feel like I've seen examples of this here, none immediately handy but will keep an eye out Gnomingstuff (talk) 05:09, 20 November 2025 (UTC)
- I know they are negative parallemisms but written in a different way but for my experience i saw the "would not be just a it would be" being used mainly by Gemini ~2025-34444-62 (talk) 03:24, 20 November 2025 (UTC)
Additional AI words
[edit]
— From this revision to Thakuri
— From this revision to Madal
— From this revision to Chinese comedy
— From this revision to Copyright law of Malaysia
— From this revision to Seoul Street Art Festival
— From this revision to Tana River County |
— From this revision to Cabaret in South Africa
— From this revision to Culture of the United Kingdom
— From this revision to Sunset Beach DJ Session 2
— From this revision to Bugis-Malay |
— From this revision to Wallingford, Connecticut
— From this revision to Bugis-Malay |
- All of these are definitely "AI vocabulary" but for this section in particular I think it is a good idea to restrict it to words where the overuse has been documented in at least one reliable source, to have something concrete to point to when people inevitably whine about it. Gnomingstuff (talk) 00:04, 25 November 2025 (UTC)
- Adding onto this, the word "noise" is used 9 times in Wikipedia:Articles for deletion/Lilly Contino, all by the same AI bludgeoner. Could it also be considered AI vocabulary? ~2025-31416-56 (talk) 14:50, 25 November 2025 (UTC)
- Have not seen any evidence in RS (or dubiously-RS) of its aggregate use, rather than in an individual case. If someone generates an AI article about Abraham Lincoln it'll almost certainly use the word "Abraham" a lot than a randomly selected human article, doesn't make that AI-specific. Gnomingstuff (talk) 17:13, 25 November 2025 (UTC)
- I think the frequent use of noise and comfort (along with discomfort and uncomfortable) may have been unique to these particular messages. LLM-generated articles normally wouldn't use words like "noise" a lot unless they're about items such as data management software designed to help businesses sort "through the noise" in order to gain insights that might help them reach and/or retain customers. Even then, the use of the word "noise" would still seem more natural (and less out-of-place) in that hypothetical AI-generated software article than in the aforementioned AfD discussion. – MrPersonHumanGuy (talk) 21:50, 25 November 2025 (UTC)
- Could "offers" (both as a noun and verb) and "opportunities" also be considered AI vocabulary? They commonly appear in spam text, and since LLMs tend to generate spammy writing, they also are overrepresented there. ~2025-31416-56 (talk) 16:55, 4 December 2025 (UTC)
- I think a few other words should be added, as it tends to overuse the words "mirror," "resonance," "ritual," and "recursive." It tends to overuse the "thread" allegory, as well as the "gilded cage." 8743ynrfy (talk) 22:42, 4 December 2025 (UTC)
- Agreed - that could be put in a subsection on AI-associated psychosis verbiage. More of this genre: I've noticed some phrases like "civilizational architecture" & an obsession with fractals and glyphs. Elysiafields (talk) 17:59, 11 December 2025 (UTC)
- On the topic of "noise", while the word "noise" appears 9 times in the deletion discussion, all of which are by the same AI bludgeoner, the last three mentions of the word are as part of the phrase "procedural noise", a phrase which appears nowhere else on the page. It's not an AI word, but this might be worth adding as a sign, in some way? ~2025-31416-56 (talk) 14:55, 11 December 2025 (UTC)
- Adding onto this, the word "noise" is used 9 times in Wikipedia:Articles for deletion/Lilly Contino, all by the same AI bludgeoner. Could it also be considered AI vocabulary? ~2025-31416-56 (talk) 14:50, 25 November 2025 (UTC)
Hallucinated core content policies
[edit]As you probably know, Wikipedia has three core content policies: verifiability, neutrality, and no original research. LLMs will usually remember the first two, but will struggle to remember the third. Rather than merely excluding it, most LLMs seem to know that Wikipedia has three core content policies, but, for some reason, cannot access no original research. These are most common in AI-generated unblock requests.
For instance, on User talk:Neutralwikifixer, an AI-generated comment identifies the core content policies as "neutrality, verifiability, and structure". Likewise, on this page, the LLM identifies Wikipedia's core content policies as "neutrality, verifiability, and reliability".
"Reliability" seems to be the LLM not understanding that it is merely a synonym for verifiability, but I have no clue where the other LLM hallucinated "structure" as a core content policy from. ~2025-31416-56 (talk) 13:26, 25 November 2025 (UTC)
- Furthermore, in the "a much more detailed prompt" section, the AI-generated comment claims to "understand the importance of original, neutral, and verifiable content". Here, as per usual, the AI understands that neutrality and verifiability are core content policies, and also includes original research. However, the LLM misunderstands the policy, and thinks that Wikipedia pages have to have original research, which is contradictory to the actual policy. ~2025-31416-56 (talk) 13:30, 25 November 2025 (UTC)
- The hallucinated "originality" as a core content policy might not be due to the AI egregiously misconstruing the no original research policy, but hallucinating that WP:PLAG is a core content policy. Since the LLM wouldn't say that it understands the importance of plagiarized content, it looks for an antonym, and eventually lands on "original". WP:NOR might have led the AI into using the word "original", however. ~2025-31416-56 (talk) 13:38, 25 November 2025 (UTC)
- In the case of biographies, the core content policies may be identified as verifiability, reliable sources and biographies of living persons. Examples from Talk:Arthur Katalayi:
- "The current revision of the article fully complies with Wikipedia’s core content policies — including WP:V (Verifiability), WP:RS (Reliable Sources), and WP:BLP (Biographies of Living Persons) — with all significant claims supported by multiple independent and reputable international sources."
- "The current version of the article strictly adheres to WP:V, WP:RS, and WP:BLP, and is backed by numerous high-quality, third-party sources from internationally recognized outlets."
- "All information in this article is fully verifiable, accurately sourced, and compliant with Wikipedia policies (WP:V, WP:RS, WP:BLP)"
- – MrPersonHumanGuy (talk) 21:33, 25 November 2025 (UTC)
- I'd hypothesize that it's because WP:NOR is not commonly mentioned in deletion discussions, at least compared to WP:BLP. ~2025-31416-56 (talk) 13:21, 26 November 2025 (UTC)
- On second thought, LLMs saying that Wikipedia has three core content policies might just be a side effect of their strict adherence to the rule of three, rather than them actually knowing that Wikipedia has three core content policies. It calling the policies "core" might just be because LLMs love using flowery language, and it's pretty much just saying "I should do these things". ~2025-31416-56 (talk) 16:54, 9 December 2025 (UTC)
Deletion discussions with major AI-generated bludgeoning
[edit]- Wikipedia:Articles for deletion/Lilly Contino
- Wikipedia:Articles for deletion/Dutch Caribbean (a large amount of the bludgeoning was removed because it was added after the discussion was closed)
- Wikipedia:Articles for deletion/Annu Gaidhu
- Wikipedia:Articles for deletion/ReggaeEDM (2nd nomination)
- Wikipedia:Articles for deletion/James Joseph Higgins
- Wikipedia:Articles for deletion/Republic of Balochistan
Most of these don't have a significantly large amount of AI-generated bludgeoning. Also, AI-generated replies are usually needlessly long. ~2025-31416-56 (talk) 18:48, 25 November 2025 (UTC)
- I think Wikipedia:Articles for deletion/Republic of Balochistan could be included on this list, as some of the bludgeoning there seemed AI-generated as well. – MrPersonHumanGuy (talk) 20:47, 25 November 2025 (UTC)
- Added. I was rushing this list, so I only included discussions where AI-generated text was collapsed. ~2025-31416-56 (talk) 13:08, 26 November 2025 (UTC)
- A pattern I've noticed with some of these is that they mention DEI quite a bit, something that doesn't apply to Wikipedia. I've never seen a human mention DEI in a deletion discussion, so it could be a sign of AI-generated writing. ~2025-31416-56 (talk) 13:17, 26 November 2025 (UTC)
- In the Annu Gaidhu deletion discussion, one LLM editor hallucinates that WP:UNDERREP is a policy, for instance. ~2025-31416-56 (talk) 13:42, 26 November 2025 (UTC)
- The editor also hallucinates WP:NOTELOCAL as a policy, presumably referring to notability from local sources. On their talk page, they claim their phone has two browsers, and then lists three. They also bring up multiple IP addresses in their unblock request, when that had nothing to do with why they were blocked.
- Might not be a sign of AI-generated writing on their own, but in the same unblock request, they pull the "the AI-generated stuff is actually mine because prompting it took a lot of effort" that a lot of people who generate AI stuff do. Lastly, they claim that they will "apply key policies," and then list a ton of policies in a parenthesis, many of which cannot be "implemented" in a traditional sense. Possibly related to false ranges, as both are LLMs stretching the meaning of words or terms. AI also loves parentheses, I've noticed. ~2025-31416-56 (talk) 15:49, 26 November 2025 (UTC)
- I don't think "mentioning DEI" is a tell.
- If you mean the literal acronym "DEI," that's only been mentioned 36 times or so by anyone in Wikipedia's history.
- If you just mean the general concepts of diversity and equity ("inclusion" is going to be hard to search for because of the deletionism vs inclusionism debate), people do and have brought that up. Doing a quick (~5 minutes) search I didn't really find much in general but there's this from 2018, this from 2007, this from 2020, etc.)
- Gnomingstuff (talk) 16:58, 26 November 2025 (UTC)
- The way I see it, it's not that LLMs are literally using the words diversity, equity and inclusion (although the user arguing to keep Anna Gaidhu mentioned "Wikipedia's equity and inclusion principles"), it's that LLMs prompted to make arguments would sometimes invoke DEI-related talking points, such as that the article on Anna Gaidhu should be kept for the sake of underrepresented and marginalized groups, or that the Dutch Caribbean article should be deleted because it reinforces "outdated colonial narratives". – MrPersonHumanGuy (talk) 12:22, 29 November 2025 (UTC)
- I agree with that. It's not a definite sign of AI, as many mentions of DEI-adjacent stuff are also from people trying to right great wrongs. However, there is a major overlap between tendentious editing and AI-generated edits. From my own experience, LLMs will become increasingly tendentious if used constantly without resetting their memory. Still, I think mentions of DEI-related talk from users are signs of AI-generated writing. While I don't doubt that there are some users who are new to Wikipedia that think that Wikipedia has some diversity and equity policy, if they start citing actual (or hallucinated) Wikipedia policies to back up their claim of diversity, then it's probably a chatbot. ~2025-31416-56 (talk) 13:26, 2 December 2025 (UTC)
- The way I see it, it's not that LLMs are literally using the words diversity, equity and inclusion (although the user arguing to keep Anna Gaidhu mentioned "Wikipedia's equity and inclusion principles"), it's that LLMs prompted to make arguments would sometimes invoke DEI-related talking points, such as that the article on Anna Gaidhu should be kept for the sake of underrepresented and marginalized groups, or that the Dutch Caribbean article should be deleted because it reinforces "outdated colonial narratives". – MrPersonHumanGuy (talk) 12:22, 29 November 2025 (UTC)
- In the Annu Gaidhu deletion discussion, one LLM editor hallucinates that WP:UNDERREP is a policy, for instance. ~2025-31416-56 (talk) 13:42, 26 November 2025 (UTC)
- A pattern I've noticed with some of these is that they mention DEI quite a bit, something that doesn't apply to Wikipedia. I've never seen a human mention DEI in a deletion discussion, so it could be a sign of AI-generated writing. ~2025-31416-56 (talk) 13:17, 26 November 2025 (UTC)
- Added. I was rushing this list, so I only included discussions where AI-generated text was collapsed. ~2025-31416-56 (talk) 13:08, 26 November 2025 (UTC)
Page too long
[edit]IMO this page has become too long and unwieldy- anybody else have the same impression ? Wuerzele (talk) 03:58, 26 November 2025 (UTC)
- I don't think it's an issue yet. Images are what slow down Wikipedia pages, and this page has only one image. My laptop with 4 gigabytes of RAM can load the page in a few seconds with no issues. ~2025-31416-56 (talk) 13:20, 26 November 2025 (UTC)
- I've been wondering about that, yeah. Not sure what to do about it as the length isn't really due to wordiness but to there being a lot of ground to cover.
- Maybe some of the entries with lots of examples can have some of them moved to "More examples" subpages (which will also allow us to theoretically have a bunch of examples in one place to better illustrate the pattern outside individual edits)? Gnomingstuff (talk) 16:45, 26 November 2025 (UTC)
Initial noodling on analyzing how AI tells change over time
[edit]I finally have a large enough dataset of AI-generated text (~1.5 million tokens) that I can kinda-sorta start to compare older AI text vs. newer AI text. I modified my AI word frequency crunching experiment:
- Split the AI article text into three folders: before August 8, 2025 (one day after the GPT-5 release), and before November 13, 2025 (one day after the GPT-5.1 release). Obviously, there is no way of knowing whether people actually used ChatGPT instead of Gemini or Perplexity or Claude or whatever, or whether they used ChatGPT Thinking, etc. ChatGPT still has the vast majority of market share so in many cases they probably did, but we don't actually know.
- Instead of comparing AI text to all human text, compare the pre-GPT-5 text to GPT-5-and-after text.
This results in about 500,000 tokens post-GPT-5 and 1 million pre-GPT-5 (still normalizing to occurrences per million). Obviously this dataset is even noisier than what we already had, and in particular the changes in frequency for individual words are currently useless. But we do start to see some interesting stuff when you get to three- and four-word phrases:
- Much more frequent before GPT-5 (bold = much much more): a testament to the, plays a crucial role, is more than just, emphasizing the need for, set the stage for, deeper understanding of, a vital role, challenges such as
- Much more frequent after GPT-5: has been profiled in, been featured in national, == Media coverage ==, == Online presence ==, cited as an example, documented in the, professional career in
Obviously, this is all very fuzzy, and I've left out a ton of flukes. But this does suggest that "undue emphasis on notability, attribution, and media coverage" is specifically a tell for newer AI text, and that a lot of the promotional AI slop phrases are dying off (which tracks with GPT-4o "glazing more"). Gnomingstuff (talk) 03:54, 27 November 2025 (UTC)
- I've definitely anecdotally observed this as of late. It's very common for edit summaries and articles to preempt criticism of a subject's notability with a statement like "They were covered in independent sources (Source1, Source2, Source3)"
- Another interesting one that seems to be on the rise is AI wikilawyering; Prolific AI user NatHaddan has a very funny section on their userpage titled "Editorial Standards" which is literally just a list of core content policies that they claim to uphold, with a statement about how their "editorial philosophy prioritises verifiability [...]"
- A brand-new user making preemptive arguments about how much they love upholding Wikipedia policy and how notable the subject of their article is is definitely a major red flag to watch. Athanelar (talk) 04:10, 27 November 2025 (UTC)
- I've also come across a few user pages that appear to follow a format involving sections titled About Me and Let's Connect! (or Let's Collaborate), and sometimes sections titled My Interests and How I Contribute in between. Some of them even use emojis. (Examples: 1 2) I'm not sure if they're pasted from an LLM or if they're all just following some sort of template. – MrPersonHumanGuy (talk) 13:07, 27 November 2025 (UTC)
Replace deleted example
[edit]A request to restore the example of unnecessary attribution from Draft:Alternative Airlines was declined; see Special:Permalink/1325022047#Draft:Alternative_Airlines. I'm looking for new examples. Pinging Graeme Bartlett (talk · contribs)... –LaundryPizza03 (dc̄) 20:21, 30 November 2025 (UTC)
- Searching draftspace for "large language model" (which hits the category of rejected drafts) is kind of fish-in-a-barrel. Unfortunately, this kind of thing tends to be pretty heavily associated with promotional content whether human or AI, so I don't know that it's going to be easy to find a non-advertising example. Gnomingstuff (talk) 21:02, 30 November 2025 (UTC)
- Bartlett suggested looking for edits to existing mainspace articles. I'm looking through Wikipedia:WikiProject AI Cleanup/Noticeboard, but haven't found anything there. Maybe also check AfD discussions with LLM hatnotes. –LaundryPizza03 (dc̄) 21:14, 30 November 2025 (UTC)
- Edits from the LLM cleanup for Bob C. Alexander (talk · contribs) look promising:
–LaundryPizza03 (dc̄) 22:25, 30 November 2025 (UTC)In the United States, university-based incubators and accelerators have expanded alongside these centers; an official Library of Congress review found that 31.5% of SBA Growth Accelerator Fund Competition winners from 2014–2016 were university-based programs.
— Special:Permalink/1317019290 - Problem solved!
–LaundryPizza03 (dc̄) 23:23, 30 November 2025 (UTC)The restaurant [The Original Pancake Kitchen] has also been mentioned in ABC News coverage relating to incidents in the surrounding precinct, underscoring its role as a well-known late-night venue in the city.
— Special:Permalink/1305163154
List of songs about Mexico
[edit]I just PROD'd the article List of songs about Mexico as an unencyclopedic, unreferenced list. If it gets deleted, here's a backup example:
List of Gokusen episodes covers episodes from the television adaptations of the Japanese manga series Gokusen.
— From this revision of List of Gokusen episodes
–LaundryPizza03 (dc̄) 09:12, 1 December 2025 (UTC)
- There's also a closely related "this article" verbiage in the same revision not noted here, as well as at this revision of First humans in Slavic mythology by the same user. –LaundryPizza03 (dc̄) 09:15, 1 December 2025 (UTC)
- The symbol between PermanentLink and the number should be /, not #. This revision is the one in which "The article surveys these narrative types..." was introduced. – MrPersonHumanGuy (talk) 14:04, 1 December 2025 (UTC)
- Thanks, corrected. –LaundryPizza03 (dc̄) 16:40, 1 December 2025 (UTC)
- The symbol between PermanentLink and the number should be /, not #. This revision is the one in which "The article surveys these narrative types..." was introduced. – MrPersonHumanGuy (talk) 14:04, 1 December 2025 (UTC)
- PROD was removed; it can probably be kept. ~2025-31416-56 (talk) 15:51, 10 December 2025 (UTC)
A thing i noticed about Gemini
[edit]Guys a stuff i realized about Gemini (mainly AI mode) is that in some alternate history scenarios it will end with "ultimately (brand name) decided to focus on its main lineup/SUVs and EVs"/the path (real life one) (the brand name) choose was the right one" (remembering i saw this mainly in alternate history scenarios) ~2025-31814-16 (talk) 15:41, 1 December 2025 (UTC)
Another example
[edit]I've noticed that some LLM edits will include notes intended for the user, but are not explicitly "collaborative communication". Should this particular example go in the Wikipedia:Signs of AI writing#Collaborative communication section, or a new subsection?
**Key changes to avoid LLM detection:**
1. **Removed promotional language**: Eliminated phrases like "comprehensive," "significant," "major," "unprecedented" 2. **Shortened lead section**: Direct, factual opening without editorial framing 3. **Simplified sentence structure**: Short, declarative sentences instead of complex clauses 4. **Removed essay-like narrative**: No "transitional period," "represents," or interpretive language 5. **Bullet points to plain lists**: More encyclopedic format 6. **Eliminated redundancy**: Removed repetitive phrases and explanations 7. **Direct factual statements**: What happened, when, where - no editorializing 8. **Stripped infobox**: Removed empty fields that looked template-generated
(To clarify, this text was added in an LLM-generated edit. For WP:BEANS purposes, that list can probably be left off, but the LLM generated the list of things that it did to avoid detection; the user then added that list into the article.) – Epicgenius (talk) 17:55, 2 December 2025 (UTC)
- Probably the latter. –LaundryPizza03 (dc̄) 06:15, 7 December 2025 (UTC)
- Dude that's actually hilarious. Guy asks it to avoid detection, then does the #1 most obvious thing to be detected.
- IMO that would go under § Collaborative communication, but you do bring up a good point with maybe not including it so we don't encourage that kind of stuff. EatingCarBatteries (contributions, talk) 06:23, 7 December 2025 (UTC)
- Pretty clear collaborative communication. If a new draft was created with something like this, it would squarely fall under WP:G15. I also don't think there's a WP:BEANS issue here, since most of what the LLM says is either insufficient for preventing detection, or requires completely rewriting the page, defeating the purpose of using an LLM in the first place. ~2025-31416-56 (talk) 15:55, 10 December 2025 (UTC)
Tagging AI-generated pages for deletion
[edit]While most AI-generated pages submitted to Wikipedia are quite obvious, WP:G15 is a somewhat narrow deletion criteria. As such, if you find a page that is almost definitely AI-generated, look to see if it qualifies for WP:G11 instead, since it doesn't require checking references and LLMs love writing in spammy language. ~2025-31416-56 (talk) 18:23, 2 December 2025 (UTC)
- Probably not a sign on its own, but a search for "Crypto" in draftspace will give a ton of AI-generated spam. ~2025-31416-56 (talk) 15:59, 3 December 2025 (UTC)
- Might as well track articles I tag as I go. This will be updated in the future. Doing this purely by tracking draftspace articles about crypto.
- G11 and G15 eligible articles: 3 (all were for reference issues)
- Probable AI-generated articles insufficient to tag for CSD: 4
- G11 eligible articles that were probably AI-generated but failed to meet G15: 3
- G15 eligible only articles: 2 (all were for use of placeholder text, 2 were duplicates)
Additional notes: AI-generated articles will commonly cite the same source under multiple names. Probable AI-generated articles insufficient to tag for CSD often are very spammy, but not spammy enough for G11. Also using abbreviations and then explaining what they mean in parentheses. If any trend can be noted, it's that the vast majority of crypto spam is AI-generated. I'll probably stop here, but I might return in the future. ~2025-31416-56 (talk) 17:31, 4 December 2025 (UTC)
- Thanks for doing this! The more data we can gather the better.
- I'd go one further -- the majority of all spam nowadays is probably AI-generated, though that doesn't mean the converse is true. Gnomingstuff (talk) 19:10, 4 December 2025 (UTC)
- G15 isn't too useful of a speedy deletion criteria. Most of the time, it takes several hours for G15-tagged pages to be deleted, since checking every reference on a page takes a large amount of time. I do think a lot of AI-generated content on Wikipedia is spam, perhaps even the majority, although that might just be due to spam being extremely common.
- For this, I was also using a very conservative definition of G11. I encountered a lot of pages that I probably could tag as G11, but I chose not to. ~2025-31416-56 (talk) 12:59, 5 December 2025 (UTC)
- Speedy deletion tags also aren't too commonly used on drafts. Simply searching for "spam" will find a good amount of G11 eligible pages because a reviewer comments that the page is spam but doesn't tag it. ~2025-31416-56 (talk) 13:02, 5 December 2025 (UTC)
- Also be careful with adding the LLM template. In many cases, it's best to just wait for the articles to be G13-deleted. ~2025-31416-56 (talk) 14:53, 5 December 2025 (UTC)
- Speedy deletion tags also aren't too commonly used on drafts. Simply searching for "spam" will find a good amount of G11 eligible pages because a reviewer comments that the page is spam but doesn't tag it. ~2025-31416-56 (talk) 13:02, 5 December 2025 (UTC)
Rule of three
[edit]"Rule of three" is listed as a sign of AI-generated writing on the page, but many human writers, including myself, tend to list things in groups of threes. While I don't doubt that something related to the rule of three is a sign of AI-generated text, merely overusing it is not. Rather, I think it should be specified to unnecessary or inappropriate use, rather than merely overuse. False ranges are a similar case.
While I don't know exactly how to rewrite it at the present moment, I do think it should be further clarified. ~2025-31416-56 (talk) 16:53, 4 December 2025 (UTC)
- AI writing absolutely overuses the rule of three because it's a convenient mechanism for generalizing vs. discussing specifics. AI generalizes due to statistical algorithms that reduce complexity to a smoothed over mean. Also since the rule of three is becoming so well known as an AI-tell, personally as a writer, I am much more careful about its usage these days. It's an old habit that at one time felt professional but now feels tainted. -- GreenC 00:49, 7 December 2025 (UTC)
Canned user pages
[edit]I've come across user pages with sections titled "About Me", "How I Contribute", "My Interests", and "Let's Connect/Collaborate!" all formatted in similar ways. – MrPersonHumanGuy (talk) 16:18, 6 December 2025 (UTC)
- i saw the link pages but the weird things is that the edit dates are from years ago (and in one page the edit date is from 2010) years before the existance of modern AI chatbots however its unknown if they are LLM generated or not ~2025-31814-16 (talk) 19:01, 6 December 2025 (UTC)
- This search returns a varied but fairly accurate sample. User pages like this are based on a template from somewhere, see this page where the template wasn't correctly populated, leaving placeholders like
[Template Contribution]
and[Language, e.g., non-English]
. fifteen thousand two hundred twenty four (talk) 20:21, 6 December 2025 (UTC)
Great New York Times article on the signs of AI writing
[edit]I'd like to share a great New York Times article that was recently published. Much shorter and more specific than this page, and pretty much summarizes and goes over the main signs of AI writing. It might be helpful to add it in as a reference or linked to as "further reading" here.
- Kriss, Sam (December 3, 2025). "Why Does A.I. Write Like … That?". The New York Times. Retrieved December 6, 2025.
BootsED (talk) 20:15, 6 December 2025 (UTC)
- I won't add in anything as it appears this project page only wants to use journal articles to describe this phenomenon, but the NYT article goes over some other tells of AI writing that don't seem to be mentioned here, such as its tendency to describe things as happening "quietly". BootsED (talk) 20:23, 6 December 2025 (UTC)
- No requirement to only use journal articles, that's just where most corroboration is coming from, and there's an off-chance people might actually believe something published in a journal. (For instance, this Pangram resource on common AI vocabulary is consistent with what I've seen in AI edits, but I haven't added it as a reference because I just know people are gonna go "well this is Pangram and Pangram is an automated AI detector and automated AI detectors will lie to you, my precious baby's life was ruined forever by false AI accusations and my mommy wrote 'intricacies' on her tombstone').
- The NYT article seems pretty good, not much new but can maybe be used to fill in some uncited blanks. Gnomingstuff (talk) 22:06, 8 December 2025 (UTC)
Loaded the text into AI and asked it to summarize below (sorry) -- GreenC 22:29, 6 December 2025 (UTC)
AI summary of NYT article
|
|---|
1. Overused Words and Vocabulary Tics
AI models tend to rely excessively on certain words that appear frequently in the "high-quality" prose they were trained on, leading to a measurable surge in their usage:
Verbs: "Delve" (especially the conjugation "delves"), "underscore," "highlight," and "showcase." The word "delves" saw a 2,700% surge in use in medical abstracts between 2022 and 2024.
Adjectives/Nouns for Complexity: "Intricate" and "tapestry" (often describing everything as a highly elaborate textile or "woven").
Adjectives for Quality/Speed: "Swift," "meticulous," and "adept."
2. Formulaic Rhetorical Patterns
AI often overfits on simple, predictable structures known to be effective in human writing, deploying them with mania:
The Em Dash (—) Obsession: The use of the em dash has become an instant tell. The article cites its frequent use in public statements like those from politicians ("not about public safety—they're about stoking fear") where it was previously reserved for more literary prose.
The Rule of Threes (Tricolons): AI text exhibits an extreme fixation on saying things in triplets ("No family. No calls. Just silence.") for a satisfying sound.
The "Not X, It's Y" Construction: The simple pattern "It's not X, it's Y" is a common tell, which sometimes morphs into the equally formulaic "No X. No Y. Just Z" in fictional narratives.
The "X with Y and Z" Tic: For lighthearted dismissals, AI frequently uses the formula: "an X with Y and Z" (e.g., calling a color "just beige with main-character syndrome and commitment issues").
3. The "Mystical Frenzy" and Abstract Sensory Language
Lacking real-world experience, AI relies on abstract associations, resulting in a distinct, often overwrought tone:
Sensory/Material Language Attached to the Immaterial: Attaching sensory qualities to abstract concepts, such as saying a day "tastes of almost-Friday," or describing "grief" or "sorrow" as having a taste or color.
Clichés and Piling Concepts: Piling concepts on top of one another until they collapse into nonsensical sentences that gesture toward deep significance but mean nothing (e.g., "Make monsters of memory. Make gods out of grief...").
4. Fiction-Specific Tropes
In AI-generated fiction, the text is marked by specific, repeated themes:
Quietness and the Spectral: An obsession with concepts like "ghosts," "shadows," "whispers," "echoes," and the word "quiet" (appearing 10 times in one 759-word essay example).
Character Names: An uncanny habit of naming female protagonists Elara Voss and male protagonists Kael in science fiction.
5. Overeager and Insipid Tone
The overall tone is often described as slightly wide-eyed, overeager, insipid, and impersonal but overwrought, reflecting an effort to reproduce the effect of "good writing" (subtlety, complexity, journey) without the genuine human experience.
|
AI creating meaningless sections
[edit]Hi all,
In AfC and NPP I have seen a lot of AI generated articles contain sections that frankly shouldn't exist, and any reasonable editor should know it's out of place. Below are the two most commons ones I've seem, but LMK if there are any others. This should probably be fleshed out more before it is given an addition
In Draft:Mojtaba Yadegari (Iranian runner), the llm created a section that specifically shows which outlets have covered him. I'm guessing the user asked the llm to display notability from multiple sources, but the llm didn't incorporate it naturally into the article, instead just showing it as a "list", fulfilling a criteria (we all know it loves lists)
Media coverage
- **IRNA** – Coverage of his inter-city marathon events.
<ref name="IRNA" />- **ISNA** – Report on an 80 km provincial peace run.
<ref name="ISNA" />- **IFRC** – Feature on his humanitarian campaigns.
<ref name="IFRC" />- **Fars News** – Interview on his national running projects.
<ref name="Fars" />- **Varzesh3** – Report on a 17-day endurance run.
<ref name="Varzesh3" />- **Borna News** – Profile on his athletic background.
<ref name="Borna" />
Another example is in Draft:Baruipur Mohun Bagan Fans' Club (BMBFC), where the llm created a section for categories. I'm guessing any reasonable person who creates an article would have read other articles, and would know that categories don't work like a See also section. On top of that, the links point to articles, not categories.
Categories
Also completely unrelated, but I've seen llms call templates that don't exist, like in Draft:Gangetic hunter-gatherers I saw {{Infobox ancient population}}. EatingCarBatteries (contributions, talk) 06:47, 7 December 2025 (UTC)
- Also, can someone tell me how I can make it so the sections that are embedded in the above quoteboxes don't show up in the TOC? EatingCarBatteries (contributions, talk) 06:51, 7 December 2025 (UTC)
- {{fake heading}} fifteen thousand two hundred twenty four (talk) 06:58, 7 December 2025 (UTC)
- Thank you, that's cool. EatingCarBatteries (contributions, talk) 07:03, 7 December 2025 (UTC)
- {{fake heading}} fifteen thousand two hundred twenty four (talk) 06:58, 7 December 2025 (UTC)
- I've seen nonexistent templates a lot, especially infoboxes, including in Draft:Answer Engine Optimization and in some LLM-generated temple articles around Wikipedia:Miscellany for deletion/Draft:Prasat Ta Leng. –LaundryPizza03 (dc̄) 06:59, 7 December 2025 (UTC)
- The comment about the mal-formatted category should be placed at WP:AIREDCAT. The other one is an interesting variety of Wikipedia:Signs_of_AI_writing#Undue_emphasis_on_notability,_attribution,_and_media_coverage. –LaundryPizza03 (dc̄) 07:30, 7 December 2025 (UTC)
overuse of "can"
[edit]Guys another thing you should add about AI chatbots is how they overuse the word "can" in texts paragraphs and sections ~2025-31814-16 (talk) 00:51, 8 December 2025 (UTC)
- Never really noticed this. Could you give examples? ~2025-31416-56 (talk) 14:30, 9 December 2025 (UTC)
- For example in my experience i saw AIS overuse the phrases:"can be seen" "can led to" "can manifest" and so many phrases with "can" on it. ~2025-31814-16 (talk) 16:33, 9 December 2025 (UTC)
- Any examples of it on Wikipedia? ~2025-31416-56 (talk) 16:38, 9 December 2025 (UTC)
- For example in my experience i saw AIS overuse the phrases:"can be seen" "can led to" "can manifest" and so many phrases with "can" on it. ~2025-31814-16 (talk) 16:33, 9 December 2025 (UTC)
Two signs not mentioned on the page
[edit]- X with Y and Z. This mainly appears when LLMs are trying to create something humorous, but it seeps into encyclopedic writing as well. Example: "Knife Guy is just a creepy name with main character syndrome and therapy needed". (not ai generated, I wrote this myself)
- Attaching sensory qualities to abstract concepts. Once again, not something that would even fit in encyclopedic text, but LLMs use it anyway. Example: "Winning had a veritable taste of glory" (once again not ai generated). LLMs feel like they're trying to attempt figurative language, but lack the emotion required to truly say anything with genuine emotional impact or meaning. Basically, imagine an alien trying to understand human emotions.
~2025-31416-56 (talk) 14:28, 9 December 2025 (UTC)
- Additionally:
- Article title as a header. Usually a level two header ==like this==. An example of the title as a level two header can be found here. Sometimes, this top header is removed, and, since LLMs understand that subheaders must be smaller, what would normally be level 2 headers are instead level 3 headers. This can be observed here.
- Bolded headers. Excessive use of boldface is already mentioned on this page, but I think that bolding of headers in particular is a sign that should be noted in the section. An example of this can be seen here.
- As well, as a sign of human writing:
- Use of quotes. From my own experience, LLMs almost never quote anything directly. As LLMs are also prone to hallucination, this is probably because direct quotes inserted by LLMs would have a large chance of being made up or approximated. Accurate quotes on a page are a major sign of it being written by a human.
- ~2025-31416-56 (talk) 14:46, 9 December 2025 (UTC)
- Mentioned this elsewhere, but if you ever need access to large volumes of AI-generated text, look here or here. There's probably a few other search terms that will give a ton of AI-generated text as results, but searching anything related to crypto is a reliable way. ~2025-31416-56 (talk) 14:50, 9 December 2025 (UTC)
- More citation signs:
- LLMs will nearly always include the site the article is about as a source, if the article is about a site. As stated earlier, LLMs do not understand that Wikipedia is not meant to include original research.
- LLMs will occasionally cite the same source under different names. This appears to be rare as I've only seen it once and cannot find it again, but it's something to keep in mind.
- ~2025-31416-56 (talk) 16:46, 9 December 2025 (UTC)
- More citation signs:
- Mentioned this elsewhere, but if you ever need access to large volumes of AI-generated text, look here or here. There's probably a few other search terms that will give a ton of AI-generated text as results, but searching anything related to crypto is a reliable way. ~2025-31416-56 (talk) 14:50, 9 December 2025 (UTC)
- Hmm not sure on some of these:
- The first two signs are more typical of blog post or fiction-style writing in my experience -- they don't really show up much in AI-generated Wikipedia-style writing since figurative language isn't really expected. If you have any counterexamples though would be interested to see, maybe on talkpage comments?
- AI actually quotes things a fair amount especially when it's being used to generate "critical reception" sections. In my experience the actual quotes tend to be accurate, but any paraphrasing of what the article says is less so.
- Using websites as a source for information on themselves isn't unheard of in human Wikipedia articles and isn't original research if it's just citing uncontroversial facts (per WP:PRIMARY).
- (if you need access to AI-generated text I'm currently at ~2 million tokens each of human vs. AI text sorted by date, not sure where one would make such a thing public though) Gnomingstuff (talk) 16:57, 9 December 2025 (UTC)
- I'm more just throwing stuff at the wall and seeing what sticks. A few more signs:
- Spaces in headers. While humans may sometimes write headers == like this ==, LLMs will almost never write headers ==like this==.
- Misplaced formatting. Far more noticeable if the LLM is also writing in Markdown (where **this** would display as bolded text, while **this **will not format properly). Should not be an indicator if the writing is done in proper MediaWiki formatting, since, as discussed on the page, things '''like '''this are more likely to be a sign of a misclick while using the visual editor.
- For signs of human writing:
- Use of HTML4 tags. Tags like <font> and <center> have been deprecated since the release of HTML5, and I have never seen LLMs use them. However, since modern browsers still support them, and they're easier to memorize than <span style="text-align: center;">, the usage of such tags is still quite common among amateurs, and HTML4 tags persist in many places on Wikipedia. Usage of such tags is probably a sign that the page was written by a human. Contrarily, the inclusion of unsupported HTML4 tags like <blink> might be a sign of AI-generated writing.
- Unclosed tags. Another sign of amateur HTML. I have never seen an LLM leave an HTML tag unclosed, but I have seen many humans do that. Likewise, missing end tags are extremely common on Wikipedia.
- ~2025-31416-56 (talk) 13:11, 10 December 2025 (UTC)
- I'm more just throwing stuff at the wall and seeing what sticks. A few more signs:
Updated data on AI vs human text
[edit]I've updated my AI word/phrase frequency experiment page with the results of comparing ~2,000,000 tokens of AI and human text each. I've also posted results for various subsets of the data:
- AI text vs. human text, restricted to the same article
- Old AI text vs new AI text
- Old AI text vs human text
- New AI text vs human text
The last three in particular may be useful for determining how many of these signs are no longer prevalent (though they may still appear in older undetected AI text) Gnomingstuff (talk) 20:34, 9 December 2025 (UTC)
- Why is Sarajevo on the list of overused 1-grams? –LaundryPizza03 (dc̄) 22:49, 9 December 2025 (UTC)
- Known fluke, one editor created a bunch of AI drafts about streets in Sarajevo. In general any proper nouns are most likely flukes Gnomingstuff (talk) 06:13, 10 December 2025 (UTC)
- Should words much more common in human writing be included as well? ~2025-31416-56 (talk) 12:55, 10 December 2025 (UTC)
- I've left the bottom half of the charts off most of those lists as they are way noisier than the AI-generated half -- a lot of proper nouns, a lot of syntactical cruft even when excluding a lot of that from the dataset. This makes sense as human writing is way more varied and less likely to coalesce around a set of words, but it's not super useful.
- For instance, of the bottom 10 three-word phrases, six of them are params like <nowiki>| doi =,and two are proper nouns like Award for Outstanding. All of those feel like flukes -- the full dataset is heavily skewed toward rejected AfC drafts, which are disproportional likely to be about non-notable people without Awards for Outstanding anything, and also skewed toward biographies, which are less likely to contain non-biography params. But they're not unambiguous flukes like, say, mentions of the year 2025, so I've been fairly conservative with what I'm excluding.
- I'm not aware of any studies that focus on words/phrases more common in human writing -- I feel like I've seen, like, Medium posts on the subject but those aren't especially reliable. If there are any then those might be worth citing if they match up. Gnomingstuff (talk) 13:58, 10 December 2025 (UTC)
- Could the human list also be focused on rejected AfC drafts? I might gather a few AI-generated drafts and human-written drafts and examine the differences. I am not a robot, so I can't do this with too many, but I think it's worth a shot for me to try.
- Then again, there is the fact that AI-generated drafts are significantly more likely to be spam... ~2025-31416-56 (talk) 14:43, 10 December 2025 (UTC)
- Here's my list:
- AI-generated:
- Draft:Alen AxP (YouTube Personality) (youtuber)
- Draft:Padi Protocol (crypto)
- Draft:Namriti (TV channel) (don't G11 this, it's gonna be G13ed soon anyways)
- Draft:Futurist concept linking demographic decline and android revolution (neologism)
- Human-written:
- Draft:Beluga (YouTuber) (youtuber)
- Draft:Oliver Carding (crypto)
- Draft:DXBM-TV (tv channel)
- Draft:Skulliness (neologism)
- I might add more later. ~2025-31416-56 (talk) 14:54, 10 December 2025 (UTC)
- For signs. Strongest indicators are bolded:
- Undue emphasis on symbolism: Present in no articles
- Undue emphasis on notability: Present in two AI-generated articles, and in one human-written article (albeit less prominently and stereotypically)
- Superficial analyses: Present in all AI-generated articles, but most prominently in the first two. Present in no human-written articles
- Promotional language: Present in three AI-generated articles, all except Draft:Futurist concept linking demographic decline and android revolution. Present in no human-written articles
- Disclaimers: Present in no articles
- Summaries: Present in no articles
- Outlines of challenges: Present in one AI-generated article and in no human-written articles
- Treating titles as proper nouns: Present in no articles
- Not doing AI vocabulary, since that's already been looked into quite a bit and it would take forever
- Negative parallelisms: Present in no articles
- Rule of three: Present in three AI-generated articles and in no human-written articles
- Vague attributions: Present in three AI-generated articles and in one human-written article
- I'm not entirely sure what elegant variation is or how to calculate it?
- False ranges: Present in no articles
- Title case headers: Present in three AI-generated articles and in no human-written articles
- Excessive use of boldface: Present in two AI-generated articles and in one human-written article
- Inline-header lists: Present in no articles
- Emojis: Present in no articles
- Overuse of em dashes: Surprisingly, present in no articles
- Cannot check for fancy quotes since the font I have installed makes them look identical
- Nothing in communication intended for the user present in any of the articles
- Markdown: Present in one AI-generated article and in no human-written articles
- Broken wikitext: Unclear. Appears to not be present in any articles
- Redlinks in categories and "see also" section: Present in two AI-generated articles and in no human-written articles
- Non-existent templates: Present in no articles
- Not doing citations since that would take forever
- Overwhelmingly verbose edit summary: Present in one AI-generated article and in no human-written articles
- Nothing else in miscellaneous
- ~2025-31416-56 (talk) 15:42, 10 December 2025 (UTC)
- Forgot to note this the first time, but the AI-generated articles without redlinks in categories and the "see also" section had no categories or "see also" section. Additionally, all AI-generated articles without a verbose edit summary had no edit summary. In all, superficial analyses the most sensitive and specific indicator that I could find, at least according to the data I gathered. ~2025-31416-56 (talk) 15:48, 10 December 2025 (UTC)
- This is purely on rejected or unsubmitted AfC submissions, and ignores unblock requests and discussions, two places where LLM use is also common and arguably an even larger issue. ~2025-31416-56 (talk) 15:50, 10 December 2025 (UTC)
- Forgot to note this the first time, but the AI-generated articles without redlinks in categories and the "see also" section had no categories or "see also" section. Additionally, all AI-generated articles without a verbose edit summary had no edit summary. In all, superficial analyses the most sensitive and specific indicator that I could find, at least according to the data I gathered. ~2025-31416-56 (talk) 15:48, 10 December 2025 (UTC)
- For signs. Strongest indicators are bolded:
- If you can find any rejected AfC drafts before ~June 2022, that'd be a great idea. Unfortunately though most drafts get deleted after 6 months which limits what's available.
- Maybe I'll look through user sandboxes -- abandoned stuff there can be deleted as of a month or so ago but most has not. Gnomingstuff (talk) 00:32, 11 December 2025 (UTC)
- OK yeah this might work, plenty of pre-2022 drafts in a search:
- Thanks for the suggestion! Gnomingstuff (talk) 00:37, 11 December 2025 (UTC)
- oops also forgot the existence of Category:Userspace drafts (which is good as I spoke too soon on sandboxes). Plenty there to work with. Gnomingstuff (talk) 05:38, 11 December 2025 (UTC)
A lot of this article is already obsolete
[edit]As we know, LLM's are still evolving rapidly. This means that the characteristics of the early versions don't necessarily apply any more. I did some quick experiments on two "signs", namely use of em-dashes for parenthetical phrases, and excessive lists of three. I asked ChatGPT (version 5.1, "thinking" mode) to write three paragraphs on random topics that I chose, aiming at "about 300 words". In the past, I would have expected at least two uses of em-dashes in each case, but only one of the three results had any em-dashes at all and then only once. In the case of lists of three, there were several but not so many and not much more than lists of 2 or 4.
After the experiment, I asked ChatGPT to explain. Regarding fewer em-dashes, it said "it’s an intentional consequence of updated training and style constraints, not just random drift—but nobody sat down and said “thou shalt stop using em dashes.” They just became less common as the models were steered toward plainer, more segmented prose."
Regarding lists of three, ChatGPT said that they are very frequent in the human-generated training data as well as rhetorically effective. Then, "Because the model is trained to imitate the distribution of patterns it sees, triads naturally pop out a lot. Newer alignment and style nudges make me slightly less repetitive about it than older versions, but it’s still baked into the statistical habits."
ChatGPT got the last word: "If you want me to use less of these characteristic signs, just ask".
It is 100% certain that programs which attempt to detect LLM writing are being used to train LLMs to not write like that. I believe that the days when LLM writing can be confidently detected are fast coming to an end. Zerotalk 07:45, 11 December 2025 (UTC)
- Old signs will just be replaced by new signs. LLMs as an underlying technology have not changed since 2017. LLMs produce patterns that are statistically common—ChatGPT said itself that
the model is trained to imitate the distribution of patterns it sees
. The part aboutNewer alignment and style nudges
refers to reinforcement learning from human feedback, which just treats the symptoms and not the underlying issues with LLMs. The fundamental limitations of LLMs, like hallucination (which isn't a bug, it's a feature) will always stay. We'll just find newer patterns to detect LLM content. - Also, it's kinda funny that ChatGPT used an em dash in an explanation of why it is using fewer em dashes. SuperPianoMan9167 (talk) 12:23, 11 December 2025 (UTC)
- LLMs now hallucinate much less than a few years ago. Humans also hallucinate (ask any teacher). There is no reason LLM hallucination can't drop to a level similar to a human. There's no reason other foibles can't also drop to a level similar to a human, and there is no reason an LLM can't imitate many more different styles than any but a very skilled human can imitate. So, yes, for a while we'll find patterns that suggest LLM but the success rate will inevitably drop off. It's a battle that will eventually be lost. Zerotalk 13:01, 11 December 2025 (UTC)
LLMs now hallucinate much less than a few years ago.
Nope.There is no reason LLM hallucination can't drop to a level similar to a human.
Actually there is. LLMs have no explicit models of the world, and they have no concept of right and wrong. Any correct answers are correct by pure chance. SuperPianoMan9167 (talk) 14:16, 11 December 2025 (UTC)- Good quote from the first link:
Wikipedia actually has a pretty good model of the world, regularly updated by its legion of editors. Every day of the week Wikipedia will correctly tell you that [Harry] Shearer was born in Los Angeles. But ChatGPT can't leverage that kind of information reliably, and doesn't have its own an explicit, stable, internal model. One day it may tell you Shearer is British, another day that he is American, not quite randomly but as an essentially unpredictable function of how subtle contextual details lead it to reconstruct all the many bits of broken-up information that it encodes.
- SuperPianoMan9167 (talk) 14:23, 11 December 2025 (UTC)
Humans also hallucinate
Not in the way that LLMs do. LLMs are essentially autocomplete on steroids: they predict the next token based on the most likely continuation of the input. What is statistically likely based on training data is not always what is correct. SuperPianoMan9167 (talk) 14:20, 11 December 2025 (UTC)- Word prediction is an important part of how LLMs produce human-like text, but miles away from a complete description of how they work. About hallucination: When I asked the first public release of ChatGPT what my birth date was, it just made up a date. If I asked 5 times, it gave 5 wrong dates. Now it searches the internet in real time and gives me the correct date. More importantly, when I asked it for the birth date of a notable person whose birth date is not on the public record, it said that it doesn't know. So it not only hallucinates less, but there are situations when it deliberately doesn't hallucinate. This, and for example the ability to solve mathematical problems it never saw before, are very hard to reconcile with "no concept of right and wrong". Zerotalk 04:35, 12 December 2025 (UTC)
- The improvements that you describe are due to two Band-Aids that are put on LLMs to reduce the symptoms: reinforcement learning from human feedback (RLHF) and retrieval-augmented generation (RAG). If you turn off the "search the web" capability (RAG), performance degrades significantly.
there are situations when it deliberately doesn't hallucinate
No, there are situations where it doesn't hallucinate based on purely random chance. The chance of hallucination may have gone down, but it will never be zero. SuperPianoMan9167 (talk) 04:40, 12 December 2025 (UTC)- I reloaded the Google search I linked above and got the correct answer, "two", exactly once out of like eight different tries. It's hit or miss. SuperPianoMan9167 (talk) 04:46, 12 December 2025 (UTC)
- Features which improve usefulness are not "band-aids", they are "enhancements". Another enhancement of ChatGPT is to execute a python program without being asked. It is only of historical interest to consider how dumb a system can be made by turning enhancements off. Humans also get dumber when their access to outside information is cut off. Zerotalk 09:04, 12 December 2025 (UTC)
- Yes, there has been improvement. That I agree with you. Whether those improvements are meaningful or not is where we disagree. SuperPianoMan9167 (talk) 13:28, 12 December 2025 (UTC)
- Features which improve usefulness are not "band-aids", they are "enhancements". Another enhancement of ChatGPT is to execute a python program without being asked. It is only of historical interest to consider how dumb a system can be made by turning enhancements off. Humans also get dumber when their access to outside information is cut off. Zerotalk 09:04, 12 December 2025 (UTC)
- I reloaded the Google search I linked above and got the correct answer, "two", exactly once out of like eight different tries. It's hit or miss. SuperPianoMan9167 (talk) 04:46, 12 December 2025 (UTC)
- Word prediction is an important part of how LLMs produce human-like text, but miles away from a complete description of how they work. About hallucination: When I asked the first public release of ChatGPT what my birth date was, it just made up a date. If I asked 5 times, it gave 5 wrong dates. Now it searches the internet in real time and gives me the correct date. More importantly, when I asked it for the birth date of a notable person whose birth date is not on the public record, it said that it doesn't know. So it not only hallucinates less, but there are situations when it deliberately doesn't hallucinate. This, and for example the ability to solve mathematical problems it never saw before, are very hard to reconcile with "no concept of right and wrong". Zerotalk 04:35, 12 December 2025 (UTC)
- LLMs now hallucinate much less than a few years ago. Humans also hallucinate (ask any teacher). There is no reason LLM hallucination can't drop to a level similar to a human. There's no reason other foibles can't also drop to a level similar to a human, and there is no reason an LLM can't imitate many more different styles than any but a very skilled human can imitate. So, yes, for a while we'll find patterns that suggest LLM but the success rate will inevitably drop off. It's a battle that will eventually be lost. Zerotalk 13:01, 11 December 2025 (UTC)
- I agree that some of these are obsolete. (See my thread above.) Doesn't mean they aren't still useful for finding pre-existing LLM edits, of which there are still a lot. It would also be really really helpful if somehow we can get information on what ChatGPT vs Claude vs Gemini vs Copilot tends to do.
- Off the top of my head:
- Undue emphasis on symbolism -- The specific "words to watch" are characteristic of 2023-2024 but the tendency is still there
- Undue emphasis on notability -- 2025 and later, this is the most easily spotted one now after markdown
- Superficial analyses -- Still appearing, especially in non-bios
- Promotional and advertising-type language: Still going strong, in part because if someone wants to spam they're probably using AI now
- Didactic/editorializing: Very 2023/early 2024, I haven't seen it much lately
- Section summaries: Very 2023/early 2024
- Outline-like summaries: May be 2024/early 2025
- List names etc treated like proper nouns -- Relatively uncommon but still happening
- Markdown -- Very much still going strong in 2025
- AI vocabulary -- Still existent but the distribution has changed. As of 2025 "delve" is dead, "underscore" is dying. Most sourcing right now has only studied up to 2024. (Note that this one is also heavily influenced by RLHF.)
- Negative parallelisms -- Still going strong but more in conversational text
- Rule of three -- Still going strong but people love to complain about this one, it is kind of an "AI uses it all the time, humans use it not infrequently" thing
- Vague attributions of opinion -- Still going strong and this is where a lot of source-to-text integrity issues sneak in -- e.g., characterizing one review of a movie as reviewer consensus
- Elegant variation -- Hard to say because a lot of editors make AI edits in chunks
- False ranges -- Probably more common in conversational text and text that is already promotional
- Gnomingstuff (talk) 14:28, 11 December 2025 (UTC)
- I would support retiring the "Didactic/editorializing" indicator. I also haven't seen them much lately and a lot of new human editors do it too (I certainly did when I was starting to edit). No longer an useful indicator. Ca talk to me! 15:13, 11 December 2025 (UTC)
- We could maybe move it to a subpage? As a first step toward (hopefully, eventually) having separate sections by timeframe. Still finding lots of undetected 2023-2024 material and this is one useful indicator for those. Gnomingstuff (talk) 18:08, 11 December 2025 (UTC)
- (hm actually, our actual indicators seem to be late 2024 and mid-2025; maybe it's just harder to find? Gnomingstuff (talk) 18:22, 11 December 2025 (UTC)
- If the amount of retired indicators starts to pile up, I'd support moving to a subpage, but for now I think having a separate section called "Obsolete indicator" would be best. Ca talk to me! 06:13, 12 December 2025 (UTC)
- We could maybe move it to a subpage? As a first step toward (hopefully, eventually) having separate sections by timeframe. Still finding lots of undetected 2023-2024 material and this is one useful indicator for those. Gnomingstuff (talk) 18:08, 11 December 2025 (UTC)
- Promotional language might be a "correlation does not equal causation" scenario. It may just be because spam text is increasingly AI-generated nowadays. While people are using LLMs to try and create legitimate Wikipedia articles, the growth of AI-generated genuine pages is significantly less than the growth of AI-generated spam. From my own experience, most LLMs won't be overly positive unless asked to be. ~2025-31416-56 (talk) 17:54, 11 December 2025 (UTC)
- I would support retiring the "Didactic/editorializing" indicator. I also haven't seen them much lately and a lot of new human editors do it too (I certainly did when I was starting to edit). No longer an useful indicator. Ca talk to me! 15:13, 11 December 2025 (UTC)
Overuse of lists
[edit]ChatGPT in particular seems to love formatting stuff in short lists. These lists usually have the title of the list item bolded and then some short clauses (usually just a single sentence) after that. These lists are also split into multiple sections. Lists seem to be mostly exclusive to GPT 5. Previous versions of ChatGPT didn't have this behavior, nor do other LLMs (including current versions).
This behavior is probably a side effect of ChatGPT attempting to truncate its output while still including all points that it deems relevant, which also explains why it doesn't happen with other LLMs. I don't have much time at the moment, but I'll add some examples later. ~2025-31416-56 (talk) 18:10, 11 December 2025 (UTC)
- Alright, I got some examples:
AVO consists of three key layers:
- SEO (Search Engine Optimization): Traditional methods for improving visibility in search engine results through content, technical, and on-page optimization.
- AEO (Answer Engine Optimization): Techniques focused on optimizing content for voice assistants and answer boxes, such as featured snippets and structured data.
- GIO (Generative Engine Optimization): Strategies for ensuring businesses are cited as credible sources in responses generated by large language models (LLMs).
— From Draft:AI Visibility Optimization (AVO). Also note rule of three.
Production Process
- The process with which a DJm composes a song generally involves the next stages:
- Concept and Lyrics — The artist defines the theme and lyrics of the song.
- AI Melodic Drafts — AI produces different melodies and rhythmic patterns following the prompt suggested by the DJm.
- Human Supervision and Enhancement — Producers adjust the instrumentation generated by the AI to match their original artistic vision.
- Layering — With the stems at hand, the DJm then combines the resulting track with new recorded pieces, including live percussion, keyboards or synthesizers.
- Mixing and Mastering — Sound balancing, effects and mastering ultimately give the song its final touch before being released. ~2025-31416-56 (talk) 13:35, 12 December 2025 (UTC)
- Given how the lists obviously appear, it's not necessary to highlight bold text like that, as that makes it rather unpleasant to look at. – MrPersonHumanGuy (talk) 14:14, 12 December 2025 (UTC)
- I've fixed it. Sorry about that. ~2025-31416-56 (talk) 16:34, 12 December 2025 (UTC)
Suspicion i have
[edit]guys its me or is this page Seal on a bedsheet AI generated since the 2 main sections cleary have these AI words superficial analysis and etc. ~2025-31814-16 (talk) 03:59, 12 December 2025 (UTC)
- @~2025-31814-16: If you think an article is AI-generated you will get a faster response posting at the dedicated AI noticeboard. SuperPianoMan9167 (talk) 04:12, 12 December 2025 (UTC)
- This talk page needs a banner that directs people to the AI noticeboard if they're looking to report suspected AI-generated content. – MrPersonHumanGuy (talk) 14:19, 12 December 2025 (UTC)
- They said they use GPT-5 for the draft on their talk page so yes Gnomingstuff (talk) 04:17, 12 December 2025 (UTC)
- The article has undergone a major edit by the Guild of Copy Editors in November 2025. I as an editor remain responsible for ensuring the accuracy and quality of any material produced using AI. The Wikipedia draft-policy on AI-generated content is a work in progress and not yet official.
- The content I have contributed does not include hallucinated information, copyright violations, claims not verified in cited sources, or fictitious references. Some notes, however, may be considered original research or unsourced (talking about the "Internationally" subsection) since I felt like the article needed a starting point for information on international examples and so I added two very general, but unsourced sentences.
- Thanks for looking out for the quality of articles on this platform. Femfem5 (talk) 09:33, 12 December 2025 (UTC)
- No problem -- I really appreciate you being transparent and providing the prompt, etc (which is more than most people do).
- Modern LLMs are better about complete hallucinations, the main place problems creep in is source-to-text integrity -- many "X highlighted Y and noted Z" statements are completely different from what they are cited to. (The AI vocabulary doesn't help here -- a common mistake is to call a throwaway one-word mention of a topic "highlighting" it.)
- That being said I didn't notice anything major when spot-checking this particular article -- there were a few "reform advocates in Iowa, Wisconsin, and Florida aimed to do X" (hypothetical example) statements when that motivation was only mentioned in the Iowa and Florida articles. But that's also an easy mistake people can make. Gnomingstuff (talk) 13:26, 12 December 2025 (UTC)
- Thanks for the response the good thing is that even if the text is AI-generated there is a lack of hallucinated information so yeah. ~2025-31814-16 (talk) 14:19, 12 December 2025 (UTC)
Broken links and big diffs
[edit]There's sections "Non-existent or out-of-place categories" and "Broken external links" but these don't really touch on internal links within the body. For example, https://en.wikipedia.org/w/index.php?title=Compute!%27s_Gazette&diff=prev&oldid=1299511754 changed a perfectly good "Commodore International" link into "Commodore 8-bit computers" (yes, the previous revision was also AI but hadn't yet screwed that up). There certainly are good cases for making red links, but this doesn't seem to be one of them. A non-LLM case where people sometimes use red links is if they are self-promoting their project, presumably in hopes someone will create a page about their game etc.
Large diffs may also be a tell, for example https://en.wikipedia.org/w/index.php?title=Draft:Commander_X16&diff=prev&oldid=1300225899 completely rewrote the article, including templates, with a woefully vague edit summary that doesn't explain the removal of images and a significant amount of writing. If you tell an LLM to make something better, it is more likely to start over than a human being would. This applies beyond Wikipedia editing, for example coding test platform Hackerrank has in its AI plagiarism detection "Suspicious Code Resetting: A large section of previously written code is deleted and replaced".
Additionally, I'm suspicious of cite errors like https://en.wikipedia.org/w/index.php?title=Google_Antigravity&diff=prev&oldid=1324473401, given Wikipedia's automatic citation tool should make things moderately foolproof.
Great article, though. This is a nightmare to try and fully cover and it's already longer than I'd like it to be. Here for the one billionth edit (talk) 01:04, 13 December 2025 (UTC)
Citation before period
[edit]@Floating Orb I do not believe this newly added indicator is an effective one. From personal experience, plenty of new human editors place citations before punctuation, likely inspired by MLA or other style guides which proscribe in-line citations before punctuation.
Mainstream AI tools also place punctuation after punctuation. In fact, ChatGPT's system prompt explicitly directs it to "Place citations at the end of the paragraph, or inline if the paragraph is long, unless the user requests specific citation placement."[1] Ca talk to me! 23:47, 15 December 2025 (UTC)
- Okay. Floating Orb Talk! my edits 23:49, 15 December 2025 (UTC)
- Courtesy explanation (high-confidence guess): By MLA, Ca may most likely be referring to the Modern Language Association, which writes this style guide. – MrPersonHumanGuy (talk) 02:02, 16 December 2025 (UTC)
Suggestion: Add a banner directing people to the AI noticeboard
[edit]When users find content somewhere on Wikipedia that they suspect of being AI-generated, they sometimes report it here and have to be reminded that the AI noticeboard is the correct place for their input. A talk page banner instructing users to report suspected AI-generated content there instead of here might save us the trouble of reminding each user. – MrPersonHumanGuy (talk) 01:56, 16 December 2025 (UTC)
Some possible deepseek syntactic quirks?
[edit]I've been editing some Deepseek text (nothing to do with Wikipedia, I've not really been active for a long time anyway), and there are two quirks that I have noticed. I don't know whether these are common amongst other chatbots, though.
- Deepseek will frequently but a comma betweeb qualifiers or subclauses ("smooth, effortless slide", "sharp, startled cry" etc.) where a human would probably do with a simple space or "and".
- It's not ungrammatical, but it gets very noticeable when it is a constant and even systematic feature of noun phrases wit two adjectives!
- It also likes to attach multiple nonfinite adjunct clauses (those centered around a gerund verb) in a way I think to be very awkward, again, usually in combination with the previous avoidance of "and".
- In fact, when I asked Gemini directly about this habit (I wasn't sure whether my ESL was leading my intuition astray), it generate this example sentence that I find incredibly bad: "Taking a deep breath, stretching her legs, looking into the distance, Jane waited for the starting gun to fire."
Circéus (talk) 03:38, 16 December 2025 (UTC)
- Might be a bit subtle/prone to misinterpretation for this page. (That's why the conjunction sign got taken off, even though the Additionally, construction is fairly high-signal for 2023-4 AI text -- too easy to misinterpret.) Gnomingstuff (talk) 15:56, 16 December 2025 (UTC)