“Archiving URLs § Preemptive Local Archiving”, Gwern2011-03-10

Archiving the Web, because nothing lasts forever: statistics, online archive services, extracting URLs automatically from browsers, and creating a daemon to regularly back up URLs to multiple sources.

In 2020-02, because of the increasing difficulty of repairing old links, I switched Gwern.net’s primary linkrot defense to preemptive local archiving: automatically mirroring locally all PDFs & web pages using manually-reviewed (and edited) SingleFile snapshots.

While it costs more time upfront (and presented some subtle UX problems like the “Arxiv problem”), it reduces total linkrot work.