7 comments

  • Antibabelic1 hour ago
    I found the page Wikipedia:Signs of AI Writing[1] very interesting and informative. It goes into a lot more detail than the typical &quot;em-dashes&quot; heuristic.<p>[1]: <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Wikipedia:Signs_of_AI_writing" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Wikipedia:Signs_of_AI_writing</a>
    • jcattle38 minutes ago
      An interesting observation from that page:<p>&quot;Thus the highly specific &quot;inventor of the first train-coupling device&quot; might become &quot;a revolutionary titan of industry.&quot; It is like shouting louder and louder that a portrait shows a uniquely important person, while the portrait itself is fading from a sharp photograph into a blurry, generic sketch. The subject becomes simultaneously less specific and more exaggerated.&quot;
      • embedding-shape24 minutes ago
        I think that&#x27;s a general guideline to identify &quot;propaganda&quot;, regardless of the source. I&#x27;ve seen people in person write such statements with their own hands&#x2F;fingers, and I know many people who speak like that (shockingly, most of them are in management).<p>Lots of those points seems to get into the same idea which seems like a good balance. It&#x27;s the language itself that is problematic, not how the text itself came to be, so makes sense to 100% target what language the text is.<p>Hopefully those guidelines make all text on Wikipedia better, not just LLM produced ones, because they seem like generally good guidelines even outside the context of LLMs.
      • robertjwebb27 minutes ago
        The funny thing about this is that this also appears in bad human writing. We would be better off if vague statements like this were eliminated altogether, or replaced with less fantastical but verifiable statements. If this means that nothing of the article is left then we have killed two birds with one stone.
      • bspammer26 minutes ago
        That sounds like Flanderization to me <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Flanderization" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Flanderization</a><p>From my experience with LLMs that&#x27;s a great observation.
      • eurekin33 minutes ago
        That&#x27;s actually putting into words, what I couldn&#x27;t, but felt similar. Spectacular quote
        • jcattle17 minutes ago
          I&#x27;m thinking quite a bit about this at the moment in the context of foundational models and their inherent (?) regression to the mean.<p>Recently there has been a big push into geospatial foundation models (e.g. Google AlphaEarth, IBM Terramind, Clay).<p>These take in vast amounts of satellite data and with the usual Autoencoder architecture try and build embedding spaces which contain meaningful semantic features.<p>The issue at the moment is that in the benchmark suites (<a href="https:&#x2F;&#x2F;github.com&#x2F;VMarsocci&#x2F;pangaea-bench" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;VMarsocci&#x2F;pangaea-bench</a>), only a few of these foundation models have recently started to surpass the basic U-Net in <i>some</i> of the tasks.<p>There&#x27;s also an observation by one of the authors of the Major-TOM model, which also provides satellite input data to train models, that the scale rule does not seem to hold for geospatial foundation models, in that more data does not seem to result in better models.<p>My (completely unsupported) theory on why that is, is that unlike writing or coding, in satellite data you are often looking for the needle in the haystack. You do not want what has been done thousands of times before and was proven to work. Segmenting out forests and water? Sure, easy. These models have seen millions of examples of forests and water. But most often we are interested in things that are much, much rarer. Flooding, Wildfire, Earthquakes, Landslides, Destroyed buildings, new Airstrips in the Amazon, etc. etc.. But as I see it, the currently used frameworks do not support that very well.<p>But I&#x27;d be curious how others see this, who might be more knowledgeable in the area.
      • andrepd13 minutes ago
        Outstanding. Praise wikipedia, despite any shortcomings wow, isn&#x27;t it such a breath of fresh air in the world of 2026.
    • paradite11 minutes ago
      Ironically this is a goldmine for AI labs and AI writer startups to do RL and fine-tuning.
  • maxbaines1 hour ago
    This is hardly surprising given - New partnerships with tech companies support Wikipedia’s sustainability. Which relies on Human content.<p><a href="https:&#x2F;&#x2F;wikimediafoundation.org&#x2F;news&#x2F;2026&#x2F;01&#x2F;15&#x2F;wikipedia-celebrates-25years&#x2F;" rel="nofollow">https:&#x2F;&#x2F;wikimediafoundation.org&#x2F;news&#x2F;2026&#x2F;01&#x2F;15&#x2F;wikipedia-ce...</a>
    • jraph1 hour ago
      I agree with the dig, although it&#x27;s worth mentioning that this AI Cleanup page&#x27;s first version was written on the 4th of December 2023.
  • feverzsj19 minutes ago
    Didn&#x27;t they just sells access to all the AI giants?
  • KolmogorovComp1 hour ago
    I wish they also spent on the reverse: automatic rephrasing of the (many) obscure and very poorly worded and&#x2F;or with no neutral tone whatsoever.<p>And I say that as a general Wikipedia fan.
    • philipwhiuk1 hour ago
      WP:BOLD and start your own project to do it.
  • progbits54 minutes ago
    The Sanderson wiki [1] has a time-travel feature where you read a snapshot just before a publication of a book, ensuring no spoilers.<p>I would like a similar pre-LLM Wikipedia snapshot. Sometimes I would prefer potentially stale or incomplete info rather than have to wade through slop.<p>1: <a href="https:&#x2F;&#x2F;coppermind.net&#x2F;wiki&#x2F;Coppermind:Welcome" rel="nofollow">https:&#x2F;&#x2F;coppermind.net&#x2F;wiki&#x2F;Coppermind:Welcome</a>
    • Antibabelic49 minutes ago
      But you can already view the past version of any page on Wikipedia. Go to the page you want to read, click &quot;View history&quot; and select any revision before 2023.
      • progbits47 minutes ago
        I know but it&#x27;s not as convenient if you have to keep scrolling through revisions.
  • weli37 minutes ago
    I don&#x27;t see how this is going to work. &#x27;It sounds like AI&#x27; is not a good metric whatsoever to remove content.
    • csande1724 minutes ago
      Wikipedia agrees: <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Wikipedia:Signs_of_AI_writing#Your_detection_ability" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Wikipedia:Signs_of_AI_writing#...</a><p>That&#x27;s why they&#x27;re cataloging specific traits that are common in AI-generated text, and only deleting if it either contains very obvious indicators that could never legitimately appear in a real article (&quot;Absolutely! Here is an article written in the style of Wikipedia:&quot;) or violates other policies (like missing or incorrect citations).
    • embedding-shape23 minutes ago
      If that&#x27;s your takeaway, you need to read the submission again, because that&#x27;s not what they&#x27;re suggesting or doing.
    • ramon15630 minutes ago
      This is about wiping unsourced and fake AI generated content, which can be confirmed by checking if the sources are valid