Internet Archive Runs 28,000 Disks on $30M/Year by Designing for Failure—Voice AI for Demos Proves the Same Pattern: Optimize for Constraints, Not Perfection

# Internet Archive Runs 28,000 Disks on $30M/Year by Designing for Failure—Voice AI for Demos Proves the Same Pattern: Optimize for Constraints, Not Perfection Internet Archive just hit HN frontpage with a deep dive into their storage architecture. **The numbers are stunning:** - **28,000 spinning hard disks** - **1.4 petabytes per rack** - **$25-30M annual budget** (entire operating budget for staff, buildings, legal, and hardware) - **No air conditioning** (uses San Francisco fog for cooling, waste heat warms building) - **"Design for failure and buy cheapest components"** (explicit engineering philosophy) For context: Storing 100 petabytes on Amazon S3 would cost **$2.1 million per month.** Internet Archive's entire annual budget is less than what AWS would charge for storage alone. The HN discussion (105 points, 20 comments, 5 hours ago) reveals the core insight: **Internet Archive succeeds by optimizing for constraints, not chasing perfection.** And this parallels Voice AI for demos perfectly: Both thrive by accepting trade-offs that make sense for their use case instead of pursuing theoretically optimal solutions that drain resources. ## The Internet Archive Philosophy: "Design for Failure and Buy the Cheapest Components" Brian Wilson, CTO of BackBlaze, articulated the principle that Internet Archive follows: > "Double the reliability is only worth 1/10th of 1 percent cost increase. The moral of the story: design for failure and buy the cheapest components you can." **Why this works:** With 28,000 spinning disks, drive failure is a **statistical certainty.** The question isn't "if" drives fail, but "how many fail per day" and "how much does it cost to fix?" **The math:** - 30,000 drives, 2% annual failure rate = 600 failed drives/year - Replacing one drive takes 15 minutes - 600 drives × 15 minutes = 150 hours of work - **Cost:** One employee for one month (~$5,000) - **Total drive cost:** $4 million **The insight:** Spending extra money to reduce failure rate from 2% to 1% saves $2,500 (half a month of salary) on a $4 million investment. **That's 0.06% savings.** It's not worth it. Buy the cheapest drives, replace them when they fail, and spend the savings on storing more data. **Voice AI for demos uses the same philosophy:** Instead of pursuing perfect DOM reading (100% accuracy on every edge case), Voice AI optimizes for: - 95%+ accuracy on common navigation patterns - Fast, lightweight implementation (doesn't require heavy ML models) - Graceful degradation (returns clear errors when element not found) **The trade-off:** Accepting 5% edge case failures saves months of development time and allows faster deployment. Users get working voice navigation today instead of perfect navigation never. ## The No-AC Insight: Use Environmental Constraints as Assets Internet Archive's Richmond District data center has **no traditional air conditioning.** **How they cool 60+ kilowatts of heat:** 1. **San Francisco fog:** Perpetual cool, damp maritime climate 2. **Ambient air circulation:** Pull in naturally cool air 3. **Higher operational temperatures:** Servers designed to run warmer 4. **Waste heat recirculation:** Capture disk heat to warm building in winter **The closed-loop efficiency:** - 60+ kilowatts of heat isn't waste—it's a resource - Winter heating comes free (byproduct of computation) - Power Usage Effectiveness (PUE) dramatically lower - Money saved on electricity → spent on more hard drives **Fallback strategies if it gets too hot:** - Delay less-urgent tasks - Reduce clock rates on some racks - Put disks into sleep mode - Power down redundant systems temporarily **The key:** Redundancy means data is available elsewhere, so temporary unavailability is acceptable. **Voice AI for demos uses the same environmental optimization:** Instead of fighting browser constraints (sandboxed JavaScript, CORS, etc.), Voice AI works with them: - DOM reading works natively in browser (no server-side rendering needed) - Accessibility tree already exists (no custom parsing required) - User permissions handled by browser (no custom auth needed) **The constraint becomes the advantage:** Browser security model prevents malicious DOM manipulation, which makes Voice AI safer by design. ## The "Data Loss Is Acceptable" Insight: Perfection Is the Enemy of Scale David Rosenthal (the article author) makes a counterintuitive point: > "We have to assume that, despite their best efforts, the Archive will over time lose some data. This isn't a problem for two reasons: > > 1. The Archive's collection is fundamentally a sample of cyberspace. Stuff missing is vastly more likely to be because it wasn't collected than because it was collected but subsequently lost. > 2. Driving the already low probability of loss down further gets exponentially more expensive. The more of the Archive's limited resources devoted to doing so, the less can be devoted to collecting and storing more stuff. Thus devoting more to eliminating data loss results in less data surviving." **The paradox:** Trying to make existing data **100% safe** means collecting **less new data**, which means **more total data loss** (from never being archived in the first place). **Better strategy:** Accept some data loss on existing content, spend resources on archiving more new content, maximize total data preserved. **Voice AI for demos faces the same trade-off:** **Chasing perfection:** - Build ML model to handle every possible DOM structure - Train on millions of websites - Achieve 99.9% navigation accuracy - Takes 18 months of development - By the time it ships, web patterns have changed **Optimizing for constraints:** - Build accessibility tree reader for common patterns - Ship working demo navigation in 3 months - Achieve 95% accuracy immediately - Use remaining time to add new features, support more use cases - Result: More users benefit sooner **The principle:** Imperfect solution shipped today beats perfect solution shipped never. ## The $25M Budget Insight: Efficiency Through Ownership, Not Rental Internet Archive's annual budget ($25-30M) is less than what AWS would charge to store their data for one year. **Why ownership scales better than rental:** **Rental model (AWS S3):** - 100 PB at $0.021/GB/month = **$2.1M/month** - Annual cost: **$25.2M** for storage alone - Plus bandwidth charges (Archive is high-traffic site) - Total: **$40M+/year** just for infrastructure - Zero ownership (vendor lock-in, price increases, terms changes) **Ownership model (PetaBox):** - Custom 4U hardware: ~$250K/rack (1.4 PB) - 100 PB ≈ 72 racks ≈ **$18M one-time** - Operational costs (power, space, maintenance): **$7-12M/year** - Total 5-year cost: **$18M + (5 × $10M) = $68M** - Full ownership (control hardware, software, no vendor dependency) **AWS 5-year cost:** ~$200M (storage + bandwidth) **PetaBox 5-year cost:** ~$68M **Savings:** **$132M over 5 years** **The trade-off:** Higher upfront investment, ongoing maintenance burden. But for long-term storage at Archive's scale, ownership wins. **Voice AI for demos makes the same ownership choice:** Instead of relying on third-party AI APIs (OpenAI, Anthropic, etc.) for every navigation command: - Build DOM reading engine once (one-time development cost) - Runs client-side (no per-request API costs) - Users control deployment (no vendor lock-in) - Scales to millions of demos without linear cost increase **The parallel:** Ownership of core technology beats rental when usage scales. ## The Kryder's Law Insight: Magnetic Storage Follows Different Economics Than Silicon David Rosenthal corrects a common misconception in the article: > "Here I have to correct Li. Moore's Law applies to silicon such as solid-state storage; it is Kryder's Law that applies to magnetic storage such as the hard disks in the PetaBox." **Why this matters:** **Moore's Law (silicon/SSD):** - Transistor density doubles every ~18 months - Performance increases exponentially - Cost per transistor drops exponentially - But: SSDs hit economic limits for bulk storage (still 5-10x more expensive than HDDs per TB) **Kryder's Law (magnetic/HDD):** - Areal density (bits per square inch) historically doubled every ~13 months - Cost per TB drops more slowly than SSDs - But: HDDs remain cheaper for bulk storage - Trade-off: Slower access, mechanical failure, but massive capacity at low cost **Internet Archive's insight:** For rarely-accessed archival data, HDD economics win. **The storage hierarchy:** - Hot data (accessed frequently): SSDs - Warm data (accessed occasionally): HDDs - Cold data (accessed rarely): Tape or deep archive Archive's data is mostly cold (old web pages rarely accessed), so HDDs are optimal. **Voice AI for demos uses the same tiering principle:** **Hot path (real-time demo navigation):** - DOM reading must be fast (accessibility tree already in memory) - User waits for immediate response - Optimize for latency, not cost **Cold path (demo analytics, historical data):** - Aggregated metrics don't need instant access - Store in database, query when needed - Optimize for cost, not latency **The pattern:** Match technology to access pattern, not to theoretical maximum performance. ## The Facebook "Cold Storage" Parallel: Tiering Based on Access Patterns The article references Facebook's 2014 cold storage architecture: > "Facebook was in the happy position of having an extremely accurate model of the access patterns to each of the small number of data types they stored, so could do a very good job of matching the data to the performance of the layer in their storage hierarchy that held it. Their expectation was that the primary reason data in the lowest layer would be accessed was that it had been subpoenaed." **Facebook's storage tiers:** 1. **Hot tier (SSDs):** Recent photos/videos, high access probability 2. **Warm tier (HDDs):** Older content, moderate access probability 3. **Cold tier (tape/optical):** Ancient content, accessed only when subpoenaed **The key:** Data automatically migrates down tiers as access probability drops. **Voice AI for demos can apply the same tiering:** **Hot tier (in-memory):** - Current DOM state - Active user session - Recent navigation history - Must be instant **Warm tier (database):** - Demo session analytics - User interaction patterns - Navigation success rates - Can take milliseconds **Cold tier (archived logs):** - Historical demo data - Long-term trend analysis - Compliance/audit logs - Can take seconds **The optimization:** Don't pay for instant access to data that's rarely accessed. ## The Petabyte-For-A-Century Problem: Proving Preservation Is Harder Than Achieving It David Rosenthal's 2006 insight reveals a fundamental problem in long-term storage: > "The basic point I was making was that even if we ignore all the evidence that we can't, and assume that we could actually build a system reliable enough to preserve a petabyte for a century, we could not prove that we had done so. No matter how easy or hard you think a problem is, if it is impossible to prove that you have solved it, skepticism about proposed solutions is inevitable." **Why this matters:** **The provability problem:** - To prove storage works for 100 years, you'd need to wait 100 years - But by then, technology has changed (can you still read the data?) - Format migration introduces new risks (each migration = chance for corruption) - Result: Perfect preservation is unprovable, so "good enough" is rational **Internet Archive's response:** - Accept some data loss - Invest in collecting more data (expand sample) - Maximize total data preserved (collection > perfection) **Voice AI for demos has the same provability challenge:** **The testing problem:** - To prove Voice AI works on all websites, you'd need to test all websites - But websites change constantly (DOM structures evolve) - New web frameworks emerge (React, Vue, Svelte, Next.js, etc.) - Result: Perfect compatibility is unprovable, so "works well enough" is rational **Voice AI's response:** - Accept some edge cases won't work - Invest in broad compatibility (accessibility tree works across frameworks) - Maximize total demos that succeed (breadth > perfection) **The pattern:** When perfection is unprovable, optimize for breadth instead. ## The 50-Petabyte Milestone: How Scale Changes Everything Internet Archive stores **100+ petabytes** of data (as of 2025). **What this means:** **At petabyte scale:** - Individual file access time doesn't matter (most data never accessed) - Individual disk reliability doesn't matter (redundancy handles failures) - Individual rack efficiency doesn't matter (aggregate efficiency matters) **What matters:** - Total cost per petabyte - Total power consumption - Total physical footprint - Total operational complexity **Internet Archive's optimizations:** - Cost: $160K to store 1 PB for 10 years (vs. millions on cloud) - Power: Natural cooling eliminates AC costs - Footprint: 1.4 PB per rack (high density) - Complexity: Simple redundancy (mirror data, no complex RAID) **Voice AI for demos operates at different scale, but same principles apply:** **At millions-of-demos scale:** - Individual demo latency matters less (most users tolerate 100ms variance) - Individual navigation accuracy matters less (95% is good enough) - Individual feature perfection matters less (core value is working navigation) **What matters:** - Total cost per demo session (must be near zero to scale) - Total deployment complexity (one-line integration) - Total maintenance burden (minimal updates required) - Total user value (does core navigation work?) **The optimization:** Ruthlessly focus on what scales, ignore what doesn't. ## The Flawed S3 Comparison: Why Apples-to-Apples Matters The original article comparing Internet Archive to Amazon S3 ($2.1M/month) is misleading because: 1. **S3 is instant-access storage** (designed for mission-critical online services) 2. **Archive's data is cold storage** (rarely accessed, long-term preservation) 3. **Bandwidth charges ignored** (S3 charges for data transfer, Archive is high-traffic) **More accurate comparison: Amazon Glacier** - Writing 1 PB to Glacier + storing 10 years + reading out: **$160K** - Still misleading (Glacier is for even colder data than Archive's) **Why comparison matters for Voice AI:** **Misleading comparison:** "GPT-4 API costs $0.03 per 1K tokens, so 1 million demo sessions × 1K tokens = $30,000/month. Voice AI must cost less." **Flawed because:** - Voice AI doesn't need LLM for every interaction (DOM reading is deterministic) - Most navigation commands are simple (no complex reasoning required) - Client-side processing eliminates most API calls **Accurate comparison:** - Voice AI demo session: 1-2 API calls for natural language parsing - Everything else: Client-side DOM reading (zero marginal cost) - Cost per million sessions: <$1,000 (vs. $30,000 for pure LLM approach) **The lesson:** Compare technologies based on actual use case, not theoretical maximums. ## The Redundancy Insight: Why Three Copies Beat One Perfect Copy Internet Archive mirrors data across multiple physical locations: - Richmond, California - Redwood City, California - Europe - Canada **Why redundancy beats perfection:** **One perfect copy:** - Must never fail (impossible) - Single point of failure (fire, flood, theft) - Requires expensive infrastructure (backup power, climate control, security) - Total cost: High **Three imperfect copies:** - Can tolerate individual failures (one copy dies, two remain) - Geographic diversity (no single disaster kills all copies) - Cheaper infrastructure per copy (don't need perfect reliability) - Total cost: Lower (3 cheap copies < 1 expensive perfect copy) **The probability math:** - Perfect copy: 99.99% uptime (best case) - Cheap copy: 95% uptime (realistic) - Three cheap copies: 1 - (0.05)³ = 99.9875% uptime **Result:** Three cheap copies are more reliable than one expensive copy, at lower total cost. **Voice AI for demos applies the same redundancy:** Instead of one perfect navigation path: - Multiple fallback strategies (accessibility tree, semantic HTML, ARIA labels, visible text) - If primary method fails, try secondary - If secondary fails, return clear error (user understands limitation) **The reliability math:** - Accessibility tree: 90% coverage - Semantic HTML: 70% coverage (for sites without accessibility tree) - Combined: 1 - (0.1 × 0.3) = 97% coverage **Result:** Multiple imperfect approaches beat one perfect approach. ## The Waste Heat Insight: Every Constraint Is an Opportunity Internet Archive turns waste heat (unavoidable byproduct of computation) into building heating (eliminates winter heating costs). **Why this matters:** **Traditional approach:** - Computation generates heat (problem) - Install AC to remove heat (cost) - Install heaters to warm building (cost) - Result: Pay twice (cooling servers + heating building) **Internet Archive approach:** - Computation generates heat (resource) - Capture heat for building warmth (savings) - Minimal cooling needed (San Francisco fog) - Result: Free heating, minimal cooling **The mindset shift:** Don't fight physics, work with it. **Voice AI for demos applies the same mindset:** **Traditional AI demo approach:** - User needs help navigating (problem) - Build complex ML model to predict intent (cost) - Train on millions of interactions (cost) - Deploy heavy model to servers (cost) - Result: Pay for complex infrastructure **Voice AI approach:** - User needs help navigating (opportunity) - DOM already contains structure (resource) - Accessibility tree already exists (resource) - Read existing data instead of predicting (savings) - Result: Lightweight architecture, client-side execution **The pattern:** Look for existing resources (DOM structure, waste heat) instead of adding new infrastructure (ML models, AC units). ## The Verdict: Optimize for Constraints, Not Perfection Internet Archive's $30M/year budget proves a fundamental thesis: **Constraints force efficient design. Efficiency scales better than perfection.** - 28,000 disks, no AC, cheap components → $68M for 5 years of 100 PB storage - AWS equivalent → $200M for 5 years - Savings → $132M that doesn't exist in Archive's budget anyway **The trade-offs Internet Archive accepts:** - Some data loss (acceptable for sampling cyberspace) - Slower access times (acceptable for archival content) - Operational complexity (acceptable for cost savings) - Higher failure rates (acceptable with redundancy) **The trade-offs Voice AI for demos accepts:** - Some edge cases fail (acceptable for 95%+ coverage) - Simpler NLU (acceptable for navigation commands) - Client-side constraints (acceptable with DOM reading) - No perfect accuracy (acceptable with clear error messages) **The pattern:** Both succeed by optimizing for constraints instead of chasing theoretical perfection that drains resources. --- **Key Takeaways:** 1. Internet Archive runs 28,000 disks on $25-30M/year by designing for failure, not perfection 2. "Buy cheapest components" beats "buy most reliable" when redundancy is cheaper than reliability 3. No AC data center uses San Francisco fog + waste heat for winter warming (environmental constraints as assets) 4. Accepting some data loss allows more total data preservation (perfection is the enemy of scale) 5. Ownership ($68M/5 years) beats rental ($200M/5 years) at Archive's scale 6. Redundancy (3 cheap copies) beats perfection (1 expensive copy) for reliability + cost 7. Voice AI for demos follows same pattern: optimize for constraints (DOM reading, client-side), not perfection (100% accuracy, server-side AI) 8. Pattern: Constraints force efficient design, efficiency scales better than perfection **Meta Description:** Internet Archive stores 100+ petabytes on $30M/year budget using 28,000 cheap disks, no AC (San Francisco fog cooling), and "design for failure" philosophy. 5-year cost: $68M vs. AWS $200M. Voice AI for demos proves same pattern—optimize for constraints (DOM reading, 95% accuracy), not perfection (ML models, 100% accuracy). Learn why efficient design scales better than perfect design. **Keywords:** Internet Archive storage architecture, PetaBox custom hardware, design for failure philosophy, cheap components redundancy, no air conditioning data center, waste heat building warming, $30 million annual budget, 28000 hard disks, Kryder's Law magnetic storage, Voice AI demos constraints optimization, client-side DOM reading, efficient architecture scales, perfection vs efficiency, long-term data preservation, ownership vs rental infrastructure
← Back to Blog