Some context about this here: https://arstechnica.com/information-technology/2023/08/openai-details-how-to-keep-chatgpt-from-gobbling-up-website-data/
the robots.txt would be updated with this entry
User-agent: GPTBot
Disallow: /
Obviously this is meaningless against non-openai scrapers or anyone who just doesn’t give a shit.
deleted by creator
If they’ll pay us when they scrape our content, sure.
… Is that like a non-argument? How do you suppose they would pay sites, let alone site users to scrape their content?
Yes that’s the point
I think this is a general question and problem for the whole fediverse, and can easily lead to the question of whether, or even when the fediverse is going to embrace having closed or private spaces or even invite only spaces, in order to try to secure some “human interaction only” social media.
That won’t stop OpenAI. We need actual blocking, on the server side. Problem is, with federation and all, it will be really, really difficult to do. And expensive.
I can understand privacy concerns, but I feel like it’s inevitable that LLMs will be used to make lots of decisions, some possibly important, so wouldn’t you want some content included in its training? For instance, would you want an LLM to be ignorant of FOSS because all the FOSS sites blocked it, and then a child asks an LLM for advice on software and gets recommended Microsoft and Apple products only?
… It’s probably going to recommend paid and non-FOSS apps and programs just on the basis that those companies probably will pay to be the top suggestions. Just like google ads. So no, I don’t think that’s a good enough reason. They can still scrape wiki’s if they need info on FOSS sites, imo. Those shouldn’t (?) block AI’s and other aggregators.
No