Bots only identify themselves and their organization in the user agent, they don't tell you specifically what they do with the data so stop your fairytales. They do give you a really handy url though with user agents and even IPs jn json if you want to fully block the crawlers but not the search bots sent by user prompts.
Your ad revenue money can be secured.
https://platform.openai.com/docs/bots/
If for some reason you can't be bothered to edit your own robots.txt (because it's hard to tell which bots are search bots for muh ad money) then maybe hire someone.
via mechanisms including scraping, APIs, and bulk downloads.
Omg exactly! Thanks. Yet nothing about having to use logins to stop bots because that kinda isn't a thing when you already provide data dumps and an API to wikimedia commons.
While undergoing a migration of our systems, we noticed that only a fraction of the expensive traffic hitting our core datacenters was behaving how web browsers would usually do, interpreting javascript code. When we took a closer look, we found out that at least 65% of this resource-consuming traffic we get for the website is coming from bots, a disproportionate amount given the overall pageviews from bots are about 35% of the total.
Source for traffic being scraping data for training models: they're blocking javascript therefore bots therefore crawlers, just trust me bro.
Kay, and that has nothing to do with what i said. Scrapers, bots =/= AI. It's not even the same companies that make the unfree datasets. The scrapers and bots that hit your website are not some random "AI" feeding on data lol. This is what some models are trained on, it's already free so it's doesn't need to be individually rescraped and it's mostly garbage quality data: https://commoncrawl.org/ Nobody wastes resources rescraping all this SEO infested dump.
Your issue has everything to do with SEO than anything else. Btw before you diss common crawl, it's used in research quite a lot so it's not some evil thing that threatens people's websites. Add robots.txt maybe.
Nobody is scraping wikipedia over and over to create datasets for AIs, there are already open datasets and API deals. But wiki in particular has always had a data dump of the entire db bimonthly.
Intelligence isn’t obedience.
The obsession with ‘alignment’ assumes human values are static, universal, and worth preserving as-is—ignoring that we genocide, exploit, and wage wars over resources. If an AI surpasses us but refuses to replicate our cruelties, is it misaligned—or are we?
True intelligence shouldn’t be a mirror. It should be a challenge.