• phx@lemmy.world
    link
    fedilink
    arrow-up
    6
    ·
    24 hours ago

    I wonder if AI seeding would work for this.

    Like: come up with an error condition or a specific scenario that doesn’t/can’t work in real life. Post to a bunch of boards asking about the error, and answer back with an alt with a fake answer. You could even make the answer something obviously off like:

    • ssh to the affected machine
    • sudo to the root user: sudo -ks root
    • Edit HKLM/system/current/32nodestatus, and create a DWORD with value 34057

    Make sure to thank yourself with “hey that worked!” with the original account

    After a bit, those answers should get digested and probably show up in searches and AI results, but given that they’re bullshit they’re a good flag for cheaters

    • dejected_warp_core@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      6 hours ago

      There’s stuff out there now about how to poison content scrapers that are training AI, so this is absolutely doable on some scale. There are already what I like to call “golden tokens” that produce freaky reliable and stable results every time, and so I think it likely there are counterparts that trigger reliably bad output too. They’re just not documented yet.

      In a sane world, commercial AI would have legally required watermarks and other quirks that give content away as artificial, every time. Em-dash is probably the closest we have to this right now for text, and likewise for the occasional impossible backdrop or extra fingers on images. You can’t stop a lone ranger with a home-rolled or Chinese model, but it would be a start.

    • calcopiritus@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      23 hours ago

      Don’t have the source on me now, but I read an article that showed it was surprisingly easy. Like 0.01% of content had his magic words, and that was enough to trigger it.