Inside The Internet Archive's Infrastructure

(hackernoon.com)

66 points | by dvrp 1 day ago

5 comments

  • hedora 19 minutes ago
    It's frustrating that there's no way for people to (selectively) mirror the Internet Archive. $25-30M per year is a lot for a non-profit, but it's nothing for government agencies, or private corporations building Gen AI models.

    I suspect having a few different teams competing (for funding) to provide mirrors would rapidly reduce the hardware cost too.

    The density + power dissipation numbers quoted are extremely poor compared to enterprise storage. Hardware costs for the enterprise systems are also well below AWS (even assuming a short 5 year depreciation cycle on the enterprise boxes). Neither this article nor the vendors publish enough pricing information to do a thorough total cost of ownership analysis, but I can imagine someone the size of IA would not be paying normal margins to their vendors.

    • toomuchtodo 13 minutes ago
      Pick the items you want to mirror and seed them via their torrent file.

      https://github.com/jjjake/internetarchive

      ia search 'format:"Archive BitTorrent"' --itemlist > itemlist.txt

      Note that there will be more than 50M items returned by this query, so that command will take a very long time to complete (results are returned in 10k chunks). You'll probably also want to add something like `--timeout 300` as well so you don't get half way through only for the command to fail with a timeout.

      https://archive.org/services/docs/api/internetarchive/cli.ht...

      u/stavros wrote a design doc for a system (Codename "Elephant") that would scale this up: https://news.ycombinator.com/item?id=45559219

      (no affiliation, I am just a rando; if you are a library, museum, or similar institution, ask IA to drop some racks at your colo for replication)

    • philipkglass 6 minutes ago
      I would like to be able to pull content out of the Wayback Machine with a proper API [1]. I'd even be willing to pay a combination of per-request and per-gigabyte fees to do it. But then I think about the Archive's special status as a non-profit library, and I'm not sure that offering paid API access (even just to cover costs) is compatible with the organization as it exists.

      [1] It looks like this might exist at some level, e.g. https://github.com/hartator/wayback-machine-downloader, but I've been trying to use this for a couple of weeks and every day I try I get a 5xx error or "connection refused".

  • cowhax 11 minutes ago
    >And the rising popularity of generative AI adds yet another unpredictable dimension to the future survival of the public domain archive.

    I'd say the nonprofit has found itself a profitable reason for its existence

  • BryantD 1 hour ago
    They have come a very long way since the late 1990s when I was working there as a sysadmin and the data center was a couple of racks plus a tape robot in a back room of the Presidio office with an alarmingly slanted floor. The tape robot vendor had to come out and recalibrate the tape drives more often than I might have wanted.
    • textfiles 25 minutes ago
      There is a fundamental resistance to tape technology that exists to this day as a result of all those troubles.
  • brcmthrowaway 1 hour ago
    Does IA do deduplication?
    • textfiles 26 minutes ago
      Not in the way I think you're talking about. The archive has always tried to maintain a situation where the racks could be pushed out of the door or picked up after being somewhere and the individual drives will contain complete versions of the items. We have definitely reached out to people who seem to be doing redundant work and ask them to stop or for permission to remove the redundant item. But that's a pretty curatorial process.
    • HumanOstrich 26 minutes ago
      Yes[1].

      [1]: The Article, Paragraph 2

  • brcmthrowaway 1 hour ago
    [flagged]