The New York Times and other major publishers have begun blocking the Internet Archive's crawlers from preserving their websites, citing concerns about AI companies scraping news content for training purposes. The Electronic Frontier Foundation argues that blocking nonprofit archivists erases decades of historical documentation and that web archiving constitutes legal fair use, similar to search engine indexing.
1 comment
The New York Times and other major publishers have begun blocking the Internet Archive's crawlers from preserving their websites, citing concerns about AI companies scraping news content for training purposes. The Electronic Frontier Foundation argues that blocking nonprofit archivists erases decades of historical documentation and that web archiving constitutes legal fair use, similar to search engine indexing.