Basically every time AI tries to create its own thing, it’s incrementally shittier than whatever it trained on. As more and more AI - produced content floods the internet, it’s increasingly training on AI - generated material. The effect is analogous to scanning and printing the same document over and over again, where it ultimately becomes a blurry mess. AI cannot create on its own, it can only modify pre-existing human work.
The article’s main solution is to keep some kind of master backup of work labelled as existing before the rise of LLMs, but isn’t optimistic of this actually happening. I’m wondering if in a few years the “write TV script” button on chatGPT generates completely unworkable garbage, will studios stop trying to pretend it’s a viable replacement for writing staff?
Calling it now, NFTs are gonna make a comeback because they’ll be used to assign provenance to human-produced work
The researchers conclude that in a future filled with gen AI tools and their content, human-created content will be even more valuable than it is today — if only as a source of pristine training data for AI.
Nonsense, we have a bright future ahead of us! As
:soypoint-1: CONTENT CREATORS
:lathe-of-heaven: :stalin-gun-1::stalin-gun-2:
Anyway what’s actually going to be used to assign provenance to human-produced work is freely-accessible video of that work being created. No-one can accuse you of using generative tools if you have video proof of your brush touching canvas.
oh shit “touch canvas” is gonna be an internet meme
Yeah but they’ll just make a plagiarism bot that fake video. They’ve already got plagiarism bots that fake the steps of drawing an image in reverse.
I think there will be an “arms race” between the generators and the verification methods, but the speedrun community for example has been dealing with this exact problem for a while and the methods of spotting fake runs are really sophisticated for the most popular games. At the very least you can ask an artist technical questions and 90% of cheaters will get weeded out because they won’t be able to talk about their process.
I’ve been fighting against AI all my life by producing only trash content.
:data-laughing:
LLMs being fed more and more generated garbage and producing increasingly worse results would be the funniest way for the AI hype to collapse.
They are discovering jpeg compression
No shit? LLMs imitate (imperfectly) human writing. A LLM trained on LLM output is going to imperfectly imitate the imperfect imitation. This is called generation loss.
https://en.wikipedia.org/wiki/Generation_loss
It’s a lossy compression thing
nods knowingly like how people in genesis lived 1000 years and now we die at 67