Comment by hyperpape
Comment by hyperpape 2 days ago
The reality is that the HTML+CSS+JS is the canonical form, because it is the form that humans consume, and at least for the time being, we're the most important consumer.
The API may be equivalent, but it is still conceptually secondary. If it went stale, readers would still see the site, and it makes sense for a scraper to follow what readers can see (or alternately to consume both, and mine both).
The author might be right to be annoyed with the scrapers for many other reasons, but I don't think this is one of them.
The reality is that the ratio of "total websites" to "websites with an API" is likely on the order of 1M:1 (a guess). From the scraper's perspective, the chances of even finding a website with an API is so low that they don't bother. Retrieving the HTML gets them 99% of what they want, and works with 100% of the websites they scrape.
Investing the effort to 1) recognize, without programmer intervention, that some random website has an API and then 2) automatically, without further programmer intervention, retrieve the website data from that API and make intelligent use of it, is just not worth it to them when retrieving the HTML just works every time.
edit: corrected inverted ratio