Comment by at_a_remove
Comment by at_a_remove 2 days ago
I have a very vague concept for this, with a different implementation.
Some, uh, sites (forums?) have content that the AI crawlers would like to consume, and, from what I have heard, the crawlers can irresponsibly hammer the traffic of said sites into oblivion.
What if, for the sites which are paywalled, the signup, which invariably comes with a long click-through EULA, had a legal trap within it, forbidding ingestion by AI models on pain of, say, owning ten percent of the company should this be violated. Make sure there is some kind of token payment to get to the content.
Then seed the site with a few instances of hapax legomenon. Trace the crawler back and get the resulting model to vomit back the originating info, as proof.
This should result in either crawlers being more respectful or the end of the hated click-through EULA. We win either way.
This doesn't work like you think it does but even if it did, do you have the money to sustain several years long legal battle against OpenAI?