Salesforce Pitches AI for Deportations, Microsoft Resurrects Failed Voice Assistant
Good Morning from San Francisco, Salesforce answered an RFP you didn't know existed. The pitch: use AI to
Wikipedia discovered an 8% decline in human visitors after improving bot detection—while the AI platforms trained on its content now answer questions without sending anyone to the source. The sustainability model breaks when citations don't include clicks.
Wikipedia revised its bot-detection systems in May, reclassified six months of traffic data, and found an 8% year-over-year decline in human visitors—as the AI models trained on its articles now answer questions directly without requiring anyone to click through to the source.
The timing compounds the problem. While human traffic falls, bot traffic explodes. Since January 2024, bandwidth consumption jumped 50%, driven by scrapers harvesting images for AI training data. When Jimmy Carter died in December, his Wikipedia page saw 2.8 million human views—manageable volume. What nearly crashed the site was automated scraping pulling a 1.5-hour debate video simultaneously. The Wikimedia Foundation now spends infrastructure dollars handling bots instead of humans, while the donation and volunteer base that funds that infrastructure depends on human visitors who increasingly don't come.
The Breakdown
• Wikipedia found an 8% year-over-year traffic decline after updating bot detection in May to catch sophisticated evasion tactics from Brazil.
• Bot bandwidth consumption jumped 50% since January 2024, with scrapers taking 65% of expensive core datacenter traffic despite being 35% of total pageviews.
• AI platforms train on Wikipedia content then answer questions directly—Pew found only 1% of users click through from Google's AI summaries.
• The sustainability model assumed citations drive traffic, traffic drives volunteers and donors—AI answers broke that link while infrastructure costs climb.
The foundation observed suspicious traffic spikes from Brazil in May 2025. Deeper analysis revealed sophisticated bots built to mimic human browsing patterns. The updated detection logic was applied retroactively to March–August 2025 data. Much of what looked like surging human interest was actually evasion bots—leaving behind an 8% real decline when the camouflage was stripped away.
The structural shift runs deeper than measurement. Google's AI overviews now answer queries with synthesized information—often from Wikipedia—without requiring a click. Pew Research found only 1% of users click through from AI summaries to source pages. Almost every major language model trains on Wikipedia datasets. Platforms extract the knowledge while eliminating the visit.
Wikipedia's openness made it essential to the internet. That same openness now threatens its survival model.
The commons that made Wikipedia valuable—free licensing, open access, volunteer curation—assumed attribution would drive traffic, and traffic would drive volunteers and donors. AI answers deliver attribution without traffic. That breaks the flywheel.
From Wikipedia's angle: We're watching value extraction without reciprocity. Platforms train models on our content, then use those models to answer questions that used to send people to our site. Infrastructure costs are climbing—50% bandwidth increase from bot scraping—while the visitor base that funds operations through donations is eroding.
From the platforms' perspective: Wikipedia is the internet's gold standard for training data—human-curated, fact-checked, comprehensive, multilingual, and legally reusable under open licensing. Why would we pay for access or redesign our interfaces to drive traffic back to the source when we can synthesize the answer directly?
From volunteers' view: A Reddit thread captures the future. "I write tutorials and project guides," one creator posted. "Google just scrapes my content and extracts snippets, and people don't visit my site anymore. There is no point in feeding AI models if I'm not going to be compensated." Wikipedia volunteers never expected compensation—just visibility, community growth, recognition. Without visitors, that motivation structure collapses.
Ad-supported publishers can chase traffic anywhere. Wikipedia's funding model assumed the citation would include the click—and AI answers severed that link.
The foundation's response reveals the bind. It's experimenting with Wikipedia content on TikTok, YouTube, Roblox, and Instagram—trying to reach younger audiences on platforms where they already spend time. But teaching people to consume Wikipedia everywhere except Wikipedia.org doesn't solve the sustainability problem. It deepens it.
The foundation's engineers discovered that 65% of their most expensive traffic—requests hitting core datacenters instead of regional caches—comes from bots. Overall bot traffic represents only 35% of total pageviews, but automated scrapers pull obscure pages and bulk downloads that can't be cached efficiently, consuming disproportionate resources.
The bandwidth chart tells the story: steady baseline demand through 2023, then a sharp inflection in January 2024 as AI training intensified. Multimedia scraping drove the 50% increase—bots harvesting the 144 million images and videos on Wikimedia Commons. When the foundation migrated systems, it found that bots made up the majority of high-cost traffic, the kind that requires forwarding requests across global infrastructure rather than serving cached copies locally.
Site reliability teams now spend significant time blocking overwhelming bot traffic before it disrupts service for human readers. The foundation is developing policies and technical frameworks for "responsible use of infrastructure"—industry language for "we need to put boundaries around automated access that's crushing our servers."
What they can't say directly: The companies scraping hardest are often the ones making Wikipedia least necessary to visit. Wikimedia Enterprise, the foundation's commercial API for large-scale reusers, was supposed to create a sustainable channel for platform access. So far, it hasn't generated enough revenue to offset the infrastructure strain or the traffic decline.
Wikipedia's sustainability rests on a three-part equation: human visitors discover content, some become volunteers who improve it, and some visitors donate to fund infrastructure. Break any leg of that stool and the structure wobbles.
Fewer visits means fewer people encounter the "edit this page" option—the entry point for the volunteer pipeline. The foundation is investing in mobile editing tools and experimenting with more "joyful" first-edit experiences, trying to convert whoever still shows up. But you can't optimize conversion rates when the top-of-funnel traffic is falling 8% year-over-year.
Individual donations fund Wikipedia's operational independence—the reason it doesn't run ads or sell access to platforms. If donor numbers follow traffic patterns, the foundation faces a choice between seeking institutional funding (which platforms could use as leverage) or curtailing services.
The foundation notes that Wikipedia "continues to remain highly trusted and valued as a neutral, accurate source of information globally." That's true for the institutions using it as training data. Less clear whether trust translates to traffic when users get Wikipedia's knowledge without visiting Wikipedia.
The downstream effects create a trap both sides would prefer to avoid. Fewer visitors erode the volunteer base. Fewer volunteers slow content updates and quality improvements. Degraded Wikipedia content reduces training data quality for the AI models that depend on it—and those models are what drove users away in the first place.
No platform has an incentive to be first-mover on sending traffic back to sources. If Google adds "click to read full Wikipedia article" prompts that users actually follow, Microsoft's Copilot captures those users with faster direct answers. If one AI chatbot requires attribution links, users switch to competitors without the friction. It's a race to extract maximum value while hoping someone else solves the sustainability problem.
The foundation's blog post appeals to users: "When you search for information online, look for citations and click through to the original source material." That's the strategic position it's stuck in—asking consumers to act against their immediate convenience to preserve a commons that serves them indirectly.
In theory, platforms depend on Wikipedia's quality and would protect their training data source. In practice, tragedies of the commons don't get solved by appeals to long-term collective interest.
The three signals that reveal whether Wikipedia's model adapts or breaks: volunteer edit rates over the next six months, individual donor trends in 2026 fundraising campaigns, and whether any major platform actually implements "click through to source" in AI-generated answers. Early read: the incentives don't favor any of those outcomes.
Q: How does Wikipedia fund itself if it's free and doesn't run ads?
A: Individual donations from users who visit the site. The Wikimedia Foundation runs periodic fundraising campaigns asking readers to contribute. That model assumes human visitors see the donation appeals—but when users get Wikipedia's content through AI answers without visiting Wikipedia.org, they never see the ask. Fewer visits means fewer potential donors.
Q: Why can't Wikipedia just block AI scrapers from accessing its content?
A: Wikipedia's entire mission rests on open access—anyone can read, reuse, and build on its content under Creative Commons licensing. Blocking automated access would contradict the founding principle that made Wikipedia valuable in the first place. The foundation wants responsible scraping with attribution that drives traffic back, but can't legally or philosophically close the gates.
Q: Why is bot traffic more expensive than human traffic if bots are only 35% of total pageviews?
A: Humans tend to read popular articles that Wikipedia caches in regional datacenters for fast delivery. Bots pull obscure pages and bulk downloads that can't be cached efficiently, forcing requests to travel to core datacenters across the globe. That's why bots generate 65% of expensive traffic despite being 35% of pageviews—they hit the costliest infrastructure patterns.
Q: Are other websites experiencing this same AI traffic problem?
A: Yes. Pew Research found only 1% of users click through from Google's AI summaries to source sites. Content publishers, libraries, and museums report similar infrastructure strain from scrapers. The difference: ad-supported publishers can chase traffic to social platforms. Wikipedia's donation model assumed people would visit Wikipedia.org specifically, making it more vulnerable when AI answers eliminate that visit.
Q: How did bots fool Wikipedia's detection systems to look human?
A: The foundation noticed suspicious traffic spikes from Brazil in May 2025 that appeared to be human browsing patterns. Investigation revealed sophisticated bots that mimicked how web browsers behave—interpreting javascript, varying request timing, and disguising automated patterns. When Wikipedia applied updated detection logic retroactively to six months of data, much of what looked like surging human interest disappeared.
Get the 5-minute Silicon Valley AI briefing, every weekday morning — free.