There’s stuff out there now about how to poison content scrapers that are training AI, so this is absolutely doable on some scale. There are already what I like to call “golden tokens” that produce freaky reliable and stable results every time, and so I think it likely there are counterparts that trigger reliably bad output too. They’re just not documented yet.
In a sane world, commercial AI would have legally required watermarks and other quirks that give content away as artificial, every time. Em-dash is probably the closest we have to this right now for text, and likewise for the occasional impossible backdrop or extra fingers on images. You can’t stop a lone ranger with a home-rolled or Chinese model, but it would be a start.
There’s stuff out there now about how to poison content scrapers that are training AI, so this is absolutely doable on some scale. There are already what I like to call “golden tokens” that produce freaky reliable and stable results every time, and so I think it likely there are counterparts that trigger reliably bad output too. They’re just not documented yet.
In a sane world, commercial AI would have legally required watermarks and other quirks that give content away as artificial, every time. Em-dash is probably the closest we have to this right now for text, and likewise for the occasional impossible backdrop or extra fingers on images. You can’t stop a lone ranger with a home-rolled or Chinese model, but it would be a start.