23 comments

  • simonw3 hours ago
    You don&#x27;t need to vandalize Wikipedia to get this kind of thing to work.<p>Back in September 2024 I named a whale &quot;Teresa T&quot; with just a blog entry and a YouTube video caption: <a href="https:&#x2F;&#x2F;simonwillison.net&#x2F;2024&#x2F;Sep&#x2F;8&#x2F;teresa-t-whale-pillar-point&#x2F;" rel="nofollow">https:&#x2F;&#x2F;simonwillison.net&#x2F;2024&#x2F;Sep&#x2F;8&#x2F;teresa-t-whale-pillar-p...</a><p>(For a few glorious weeks if you asked any search-enabled LLM, including Google search previews, for the name of the whale in the Half Moon Bay harbor it confidently replied Teresa T)
    • sb0571 minute ago
      I mean, the name of that whale is now Teresa T. You gave it that name.
    • pesus21 minutes ago
      Google <i>still</i> shows Theresa T as the name when you search.
    • slater2 hours ago
      (it probably helps that your name &amp; blog carry some weight, vs. some rando writing something on blogspot or wordpress ;) )
      • Forgeties792 hours ago
        Which illustrates another problem: unscrupulous actors with big names can spread whatever information they want to millions of people with minimal effort.
        • simonw1 hour ago
          Exactly. I chose to abuse my platform to promote Teresa T as the name of a whale.
          • Forgeties7929 minutes ago
            Oh god I just realized the implication! I was not directing that at you haha
        • MassPikeMike1 hour ago
          Ever since the invention of the printing press, every new communication technology has reduced the effort needed to widely disseminate information-- and misinformation! So you could say this is nothing new. On the other hand, this is remarkably little effort.
        • nomdep1 hour ago
          Yes, they can. We can be glad that respectable newspapers and TV news channels have never done it and never will. You can even trust than the headlines are accurate summaries of the content of the articles. &#x2F;s
    • bitwize2 hours ago
      The Mr. Splashy Pants of the AI era!
  • nicole_express1 hour ago
    It&#x27;s an odd thing here, because I don&#x27;t really understand why this is LLM-specific at all. If someone came up to me and asked &quot;who&#x27;s the 6 Nimmt world champion?&quot; I&#x27;d google it and probably find the same result, and have no reason not to believe it. I mean, for all I know the game is being made up too, though it has more sources at least.
    • yen22337 minutes ago
      A lot of people seem to think this to be an LLM problem, but you&#x27;re right.<p>This is a general epistemological problem with relying on the Internet (or really, any piece of literature) as a source of truth
    • refulgentis54 minutes ago
      Closed it after “This house of cards only needs a $12 domain!”, right under “Sorry, Wikipedia”, right under their Wikipedia edit.
  • xeeeeeeeeeeenu2 hours ago
    The key to successful poisoning attacks is to introduce brand new information that doesn&#x27;t directly contradict other training data. It&#x27;s much easier to convince the LLMs that you&#x27;re the king of a fictional Mapupu kingdom than the president of the United States.<p>So this means that for bad actors it&#x27;s more efficient to manufacture brand new fake stories instead of trying to distort the real ones. Don&#x27;t produce fake articles absolving yourself of a crime, instead produce fake articles accusing your opponent of 100 different things. Then people will fact-check the accusations using LLMs, and since all the sources mentioning those accusations are controlled by you, the LLMs will confirm them.
  • billypilgrim2 hours ago
    I must say I expected an actual poisoning of the data used to train the LLM and was excited, but the examples indicate that the LLM just searched the web and reported what it found? When you create a website with fake information and search Google for that information, it will of course bring up your site, not because it’s factually correct but because it’s related to what you searched for. What am I missing?
    • rincebrain1 hour ago
      The part where lots of people have historically trusted LLM responses without verification, more than trying to sort through the dross on Google or Bing search results is, I think, the point.
  • blobbers2 hours ago
    This is basically the same problem of products astroturfing reddit, or SEO optimizing google. You want a new X, and so they heavily go after the keywords associated with it.<p>This is sort of why &quot;brand&quot; matters; it provides a source of trust.<p>Encyclopedia Britannica used to be that source of &#x27;facts&#x27;. Then it became whatever page-rank told you. Eventually SEO optimization ruined that.<p>News stories are the same thing. For certain groups, they have their &#x27;independent&#x27; publication whose reporting they trust.
    • nailer2 hours ago
      It&#x27;s such a pity the Oxford English Dictionary decided to paywall themselves decades ago - they used to be THE dictionary in most countries, now nobody seems to know who they are.
      • blobbers1 hour ago
        They would have been better off going freemium or ad-supported. Or 501(c)(3) ala wikipedia?
      • anikom151 hour ago
        The OED’s goal isn’t really to be every nation’s dictionary.
  • _carbyau_43 minutes ago
    One of the problems with labelling automation as AI.<p>People think that whatever information an &quot;AI&quot; spits out has gone through a round of critical thinking which enhances the trust value of that information.<p>The early LLM&#x27;s using groomed data <i>may</i> have had such critical thinking somewhere in the pipeline. So it was already not really trustworthy.<p>And now? Using agents to search the internet for you?...<p>Garbage in, garbage out still applies in computing as ever.
  • amarant3 hours ago
    &quot;Stoner became the first American world champion....&quot;<p>Even being on stoner.com,I read that as meaning something different from what was meant.<p>Op has a great surname!
  • yen22343 minutes ago
    I feel uncomfortable that I can&#x27;t actually verify that this story is true.<p>Asking Opus 4.7 who the reigning 6nimmt! champion is leads to this article and a warning about a possible hoax
  • jrmg2 hours ago
    BBC journalist doing a very similar thing in February: <a href="https:&#x2F;&#x2F;www.bbc.com&#x2F;future&#x2F;article&#x2F;20260218-i-hacked&#x2F;-chatgpt-and-googles-ai-and-it-only-took-20-minutes" rel="nofollow">https:&#x2F;&#x2F;www.bbc.com&#x2F;future&#x2F;article&#x2F;20260218-i-hacked&#x2F;-chatgp...</a>
  • Paracompact2 hours ago
    Most of the popular discourse around AI is still at the level of, &quot;Don&#x27;t trust the AI, trust the sources!&quot; When it gets to the point where even the sources of simple facts are untrustworthy, the average person just trying to learn some trivia about the world is doomed.<p>Doesn&#x27;t help that AI media literacy is so primitive compared to how intelligent the models are generally. We&#x27;re in a marginally better place than we were back when chatbots didn&#x27;t cite anything at all, but duplicated Wikipedia citations back to a single source about a supposedly global event is just embarrassing. By default, I feel citations and epistemological qualifications should be explicit, front-and-center, and subject to introspection, not implicit and confined to tiny little opaque buttons as an afterthought.
    • amiga3862 hours ago
      Wikipedia calls this <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Citogenesis" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Citogenesis</a> (after XKCD coined it).<p>You can expect the spicy autocomplete to feed you flattering bullshit. It may cite Wikipedia (it shouldn&#x27;t), but you should go check out those citations, and validate the claims yourself. It&#x27;s the least you can do.<p>And if the cited source is Wikipedia... check <i>Wikipedia&#x27;s</i> sources too. Wikipedians try their best to provide you with reliable sources for the claims in their articles (oh who am I trying to kid? They pick their favourite sources that affirm their beliefs, and contending editors remove them for no good reason, and eventually the only thing that accrues is things that the factions agree on, or at least what ArbCom has demanded they stop fighting over).<p>I guess what I&#x27;m trying to say is: don&#x27;t rely on that authoritative-sounding tone that Wikipedia uses (or that AI bots use, or that I&#x27;m using right now). It&#x27;s a rhetorical trick that short-circuits your reasoning. Verify claims with care.<p>Also check the Talk page, you often find all kinds of shenanigans called out there.
      • bitwize2 hours ago
        Perhaps my favorite example of a citogenesis-like process is the legendary arcade game Polybius, which originated as an entry on some German guy&#x27;s web compendium of arcade games (coinop.org), perhaps as a &quot;paper town&quot;, or fake entry that acts as a copyright canary when duplicated elsewhere. Gamer news and special-interest blogs and sites, and even print publications like <i>GamePro</i> picked it up, and I think it was even listed on Wikipedia as an urban legend whose actual existence was unknown. Then the retrogaming YouTuber Ahoy did an in-depth documentary (<a href="https:&#x2F;&#x2F;m.youtube.com&#x2F;watch?v=_7X6Yeydgyg" rel="nofollow">https:&#x2F;&#x2F;m.youtube.com&#x2F;watch?v=_7X6Yeydgyg</a>) which concluded that Polybius didn&#x27;t exist and was never even mentioned before the aforementioned coinop.org reference and, for me anyway, that settled it. Polybius, in its urban legend form, never existed.<p>(Norm Macdonald voice) Or so the Germans would have us believe...!
        • egypturnash11 minutes ago
          And then an insane Welsh game wizard made it real. <a href="http:&#x2F;&#x2F;minotaurproject.co.uk&#x2F;Virtual&#x2F;Polybius.php" rel="nofollow">http:&#x2F;&#x2F;minotaurproject.co.uk&#x2F;Virtual&#x2F;Polybius.php</a>
  • duxup1 hour ago
    In American college football there&#x27;s all sorts of awards, and each year they put out &quot;watch-lists&quot; and silly press releases that get parroted on social media by any team that has their own player mentioned.<p>I&#x27;ve wanted to come up with my own for a while ...
  • Lerc1 hour ago
    How many people have done things like this and then disclosed the fact? It would be fascinating to collect as many instances as you can to develop a data set. Could you train a system to find more? How many could it find, and in what areas?
  • gverrilla1 hour ago
    Poisoning wikipedia shows low respect.
  • drchiu2 hours ago
    My wife cited ChatGPT as her primary source the other day when she wanted to debate with me on something.<p>&quot;AI told me that...&quot;<p>In the old days, it would have been &quot;I read on Google...&quot;
  • CrzyLngPwd2 hours ago
    So it&#x27;s trivial for an individual to poison the LLMs, but imagine what a state with billions of American dollars could achieve.<p>We can easily look ahead a few years and see how people will rely on the LLMs to be a source of truth in the same way people looked at Google that way, or newspapers.<p>Rewriting history has been happening for a while, and with LLMs being the one-stop shop for guidance and truth, the rewrite will be complete.<p>Doubly so since most people see these things as artificial intelligence, and soon to be superintelligence...so how can they be wrong?
  • standeven3 hours ago
    I&#x27;ve had LLMs regurgitate satire as fact many, many times.
  • poglet1 hour ago
    I made a post on Reddit asking for help with a TV, I had made up some (likley incorrect) technical assumptions about the issue. Several years later I asked the LLM about the TV, it used my own post as a citation to tell me what was wrong with it.<p>I am paranoid that this is happening every time I ask a LLM for a product recommendation or a shop recommendation. In the same way as SEO, anyone wanting to sell or convince needs to do as much as they can to influence the LLM.
  • Havoc2 hours ago
    Like a FIFA peace prize?
  • shevy-java2 hours ago
    So like Frank Dux! In the movie Bloodsport epilogue, he didn&#x27;t do that.<p>It&#x27;s almost like he was a better Chuck Norris than Chuck Norris. By his own ... testimony ...
  • nonameiguess2 hours ago
    Pails in comparison to what Frank Dux and Frank Abagnale were able to convince much of the world they did with no evidence other than their own stories. Who knows how much of recorded and believed history is complete bullshit? Not to get too far into sacred territory, but claims around Siddhartha Gautama, Jesus Christ, and the Prophet Muhammad are quite a bit less plausible than the legends of Ragnar Lodbrok or the tales of Jonathan Swift, but nonetheless widely believed.
  • blobbers2 hours ago
    [dead]
  • nailer2 hours ago
    [flagged]
  • dyauspitr3 hours ago
    Why does this person deserve any kind of support? What’s the point of poisoning LLMs? To put some cursory Luddite roadblock that might delay the technology for a couple of months?
    • jurgenkesker3 hours ago
      Support? It&#x27;s just showing weaknesses of LLM&#x27;s. Which is a valid sort of research I would say?
      • wewtyflakes2 hours ago
        That&#x27;s fair, though on the other hand it kind of feels like &quot;Don&#x27;t drive cars, there could be rocks on the road! See, just look at all these rocks I put on the road!&quot;. Which is true, and real, but perhaps frustrating for people who just want to get someplace in peace.
    • alnwlsn1 hour ago
      To prove you can. Which means someone else with more to gain from it will probably do it also, and you should probably expect this to happen.
    • jrmg2 hours ago
      This is a “if we stopped testing there would be far fewer cases!” mentality...
    • duskwuff2 hours ago
      &gt; What’s the point of poisoning LLMs?<p>It&#x27;s a demonstration. If a domain name and a quick bit of Wikipedia vandalism is all it takes to make an LLM start spouting nonsense about a &quot;surprisingly serious tournament circuit&quot; or a &quot;massive online community&quot; for an obscure card game, consider what an unscrupulous PR team or a political operative could do to influence its output on more important topics.
      • nickthegreek2 hours ago
        &gt; consider what an unscrupulous PR team or a political operative could do to influence its output on more important topics.<p>‘is doing’.
    • ethin3 hours ago
      You do know that calling people who don&#x27;t like AI for any reason Luddites does you no favors, right? It just makes you look like your a part of a cult.