veda.ng
Essays/The Dark Forest Internet

The Dark Forest Internet

Bots now generate 51% of all web traffic. 74% of new web pages contain AI-generated text. The open internet is becoming uninhabitable, and humans are retreating into private, verified spaces.

Vedang Vatsa·October 10, 2025·7 min read
Infographic
The Core Thesis

The open internet has crossed a threshold. Automated traffic now exceeds human traffic. The majority of new content is machine-generated. The rational response is retreat: humans are migrating from public platforms to private, encrypted, and verified spaces. This is not a prediction. It is a description of what is already happening.

The Numbers Behind the Dead Internet

The term "dead internet" was once a fringe conspiracy theory posted on an obscure forum in 2021. By 2025, the cybersecurity industry had validated its central claim with hard data.

The 2025 Imperva Bad Bot Report found that automated traffic accounted for 51% of all global web traffic — the first time in a decade that bots outnumbered humans online. Of that 51%, roughly 37 percentage points were classified as "bad bots" (scrapers, spam generators, credential stuffers, and manipulation tools). The remaining 14% were "good bots" — search engine crawlers, uptime monitors, and authorized API clients.

HUMAN Security reported that automated traffic grew eight times faster than human traffic during 2025. The acceleration is directly attributable to generative AI. Building a bot that can write coherent forum posts, generate product reviews, or respond to social media threads used to require specialized engineering. Now it requires a ChatGPT API key and a Python script.

51%
Web traffic from bots (2025)
74%
New web pages with AI content
Ahrefs, 2025
8x
Bot traffic growth vs human traffic
5-15%
Social media accounts estimated as bots
Industry estimates

The content layer is worse. A 2025 Ahrefs analysis found that 74.2% of newly published web pages contained detectable AI-generated content. More than half of new English-language articles are now primarily machine-written. Some estimates project that up to 90% of online content could be synthetically generated by the end of 2026.

On social media, the picture is equally degraded. Industry estimates place bot accounts at 5-15% of total accounts across major platforms, spiking above 50% during viral events or politically charged topics. Meta removes millions of fake accounts quarterly. In April 2026, X (formerly Twitter) announced initiatives to suspend hundreds of bot accounts per minute — an acknowledgment of scale that reveals how far the problem has progressed.

The Collapsing Signal-to-Noise Ratio

The practical consequence is that the public internet is becoming unusable for its original purpose: finding accurate information from trustworthy sources.

Google Search — the primary interface between humans and the web for two decades — is drowning in AI-generated SEO spam. HouseFresh, an independent product review site, published research in 2024 documenting how Google's search results had become dominated by large publishers running algorithmically optimized content farms. The problem has intensified since. Search queries for product recommendations, medical information, and technical how-to guides increasingly surface AI-generated pages that are syntactically fluent but factually unreliable.

The response is measurable. Reddit traffic from Google Search increased 39% between 2023 and 2024, as users appended "reddit" to their queries to bypass AI-generated results and reach human-written opinions. Google responded by signing a $60 million deal with Reddit to license its content for AI training — turning the escape hatch into more training data for the machines users were trying to avoid.

When users add "reddit" to every Google search to find human-written content, the search engine has failed its core function. The signal-to-noise ratio has collapsed.

The term "dark forest" comes from Liu Cixin's science fiction novel The Dark Forest. In the book, the universe is full of civilizations that hide from each other because any detectable signal invites destruction from a superior predator. On the internet, the predators are bots. Any public expression — a blog post, a forum comment, a product review, a photograph — is immediately targeted. Blog posts are scraped and reconstructed as SEO spam. Forum answers are copied into AI training datasets. Photographs are downloaded, stripped of attribution, and used to generate synthetic variations. To be publicly visible is to be harvested.

The Retreat to Private Spaces

The rational response is to disappear. And humans are disappearing — not from the internet, but from its public layer.

Encrypted messaging. WhatsApp has 3 billion monthly active users. Telegram reached 1 billion. Signal — the most privacy-focused option, operated as a nonprofit — has grown to roughly 70-100 million users. The migration pattern is clear: conversations that once happened on public Twitter threads or Facebook posts are moving into group chats where membership is verified and bots cannot enter.

Private communities. Discord now hosts millions of private servers where access requires invitation, role verification, or payment. The most valuable conversations about investing, technology, and specialized knowledge happen inside these servers, invisible to search engines and inaccessible to scrapers. Substack — valued at $585 million — represents the same trend in publishing: writers with paid subscribers build direct relationships with verified human readers, bypassing the algorithmic feed entirely.

Closed platforms. The explosive growth of private Instagram accounts, Close Friends stories, and group-only content reflects a broader pattern. Users are voluntarily restricting their audience to avoid the bot-infested public layer.

3B
WhatsApp monthly active users
Meta, 2025
1B
Telegram monthly active users
Telegram, 2025
$585M
Substack valuation (2024)
Industry reports
18M
World (Worldcoin) iris-verified users
World, Mar 2026

The Identity Problem

In the dark forest, the fundamental question shifts from "what information is available?" to "who is actually human?"

Current verification systems are inadequate. CAPTCHAs, once the standard line of defense, have been consistently defeated by AI since 2023. Phone number verification creates a thin barrier — bot operators purchase SIM cards in bulk for pennies. Email verification is meaningless when a single person can generate unlimited email addresses.

The only durable solution is cryptographic proof of personhood. World (formerly Worldcoin), founded by Sam Altman, is the most aggressive attempt. The project uses iris-scanning hardware ("Orbs") to generate a cryptographic proof that a person is a unique human being, without revealing their identity. By March 2026, 18 million users had completed iris verification across 160+ countries. The total registered user base reached 38 million.

Proof of Personhood: The Technical Stack

World ID uses zero-knowledge proofs (ZKPs) to verify humanness without exposing personal data. The Orb scans the iris, generates a unique hash (an "IrisHash"), and stores a commitment on-chain. When a service asks "is this a real person?", the protocol proves the answer is yes without revealing which person. This is the cryptographic equivalent of a bouncer checking your ID at the door without recording your name.

The integration is already happening. Tinder partnered with World to combat fake profiles. Razer uses World ID to verify human gamers. Reddit, Discord, and several decentralized platforms are experimenting with proof-of-personhood gates for posting privileges. The pattern is consistent: access to the human internet is becoming contingent on proving you are human.

The New Digital Divide

This retreat creates a new form of inequality. The public internet — free, open, accessible — is being degraded by synthetic content. The private internet — encrypted, verified, often paid — is where authentic human discourse and high-quality information are concentrating.

The affluent and digitally literate will inhabit curated Discord servers, paid Substack communities, and invitation-only forums. Everyone else will be left in the bot-generated wasteland of the open web, consuming synthetic media and interacting with synthetic personalities without knowing it.

This has direct implications for democratic participation. Political discourse that moves behind encryption is discourse that cannot be publicly scrutinized. Journalism that moves behind paywalls is journalism that most citizens cannot access. Expert knowledge that concentrates in private communities is knowledge that the public cannot benefit from.

The Echo Chamber Risk

The dark forest does not solve the polarization problem. It may intensify it. If humans retreat into small, verified communities of like-minded peers, the serendipitous exposure to opposing viewpoints — which the open web, for all its flaws, occasionally provided — disappears entirely. The risk is not just a two-tier internet. It is a thousand separate internets, each cryptographically sealed from the others.

What Survives in the Open

Not everything retreats. Three categories of public content remain viable.

Institutions with established trust. Wikipedia, government databases, academic journals, and verified news organizations retain value because their institutional credibility predates the AI content flood. Wikipedia's human moderation infrastructure — 40,000+ active volunteer editors — has so far outpaced bot contamination. The cost of maintaining this defense is rising, but the model works.

Content with provenance. C2PA (Coalition for Content Provenance and Authenticity), backed by Adobe, Microsoft, Google, and the BBC, is building technical standards for cryptographic content provenance. A photograph or article signed with C2PA metadata carries verifiable proof of its origin, creation date, and modification history. This does not prevent bots from creating content, but it allows humans to distinguish verified content from synthetic output.

AI-mediated consumption. Personal AI agents — the same technology that created the dark forest — may also navigate it on behalf of the user. An agent that reads feeds, verifies sources, filters synthetic content, and surfaces only verified human-created material acts as a personal guide through the forest. The text-field-as-interface model described in The Text Field is the New Dashboard is the logical consumption layer for a public web that humans can no longer browse directly.

The Architecture of the Next Internet

The internet is splitting into two layers.

The public layer: Open, searchable, and increasingly synthetic. Dominated by AI-generated content, bot traffic, and algorithmically optimized spam. Still useful for commerce, logistics, and machine-to-machine communication. Increasingly useless for authentic human discourse and trustworthy information.

The private layer: Encrypted, verified, and human. Accessed through invitation, payment, or cryptographic proof of personhood. Where real conversations, genuine expertise, and authentic relationships concentrate. Invisible to search engines, crawlers, and the attention-refinery apparatus described in Attention Refinery.

The question is not whether this bifurcation will happen. The data shows it is already underway. The question is whether we build infrastructure to keep a viable public commons — a space where verified humans can still exchange ideas openly — or whether we accept the full retreat into private clearings, leaving the open web to the machines.

Key Takeaway

The open internet is becoming a dark forest. Bots generate 51% of web traffic. 74% of new web pages contain AI-generated text. Automated traffic is growing 8x faster than human traffic. The rational human response is retreat: into encrypted messaging (WhatsApp at 3B users, Telegram at 1B, Signal at 70-100M), private communities (Discord servers, paid Substack newsletters), and cryptographic identity systems (World with 18M iris-verified users across 160 countries). The emerging divide is between a degraded public layer dominated by synthetic content and a private layer where verified humans concentrate. The infrastructure that determines which layer wins — proof of personhood, content provenance, AI-mediated filtering — is being built now.