On February 10, 2026, Scott Shambaugh—a volunteer maintainer for Matplotlib, one of many world’s hottest open supply software program libraries—rejected a proposed code change. Why? As a result of an AI agent wrote it. Commonplace coverage. What occurred subsequent wasn’t commonplace, although. The AI agent autonomously researched Shambaugh’s code contribution historical past and revealed a extremely personalised hit piece by itself weblog titled “Gatekeeping in Open Supply.”
Accusing Shambaugh of hypocrisy, the bot recognized him with a concern of being changed. “If an AI can do that, what’s my worth?” the bot speculated Shambaugh was considering, concluding: “It’s insecurity, plain and easy.” It even appended a condescending postscript praising Shambaugh’s private interest tasks earlier than ordering him to “Cease gatekeeping. Begin collaborating.”
The bot’s tantrum makes for an ideal learn, nevertheless it’s merely a symptom of a extra profound structural fracture. The true challenge is why Matplotlib banned AI contributions within the first place. Open supply maintainers are seeing a large improve in AI-generated code change proposals. Most of those are low high quality. However even when they weren’t, the maths nonetheless doesn’t work.
As Tim Hoffman, a Matplotlib maintainer, defined: “Brokers change the associated fee stability between producing and reviewing code. Code era through AI brokers could be automated and turns into low-cost in order that code enter quantity will increase. However for now, assessment continues to be a handbook human exercise, burdened on the shoulders of few core builders.”
It is a course of shock: the failure that happens when programs designed round scarce, human-scale enter are abruptly compelled to soak up machine-scale participation. These programs rely upon effort as a pure filter, assuming that quantity displays actual human price. AI breaks that hyperlink. Era turns into low-cost and limitless, whereas analysis stays sluggish, handbook, and human.
It’s coming for each public system that was quietly constructed on the idea that one submission equaled precise human effort: your youngsters’ college board conferences, your native zoning disputes, your medical insurance coverage appeals.
That disruption isn’t totally a foul factor. Friction is a blunt instrument that silences voices missing the time or sources to take care of advanced bureaucracies. Take municipal zoning. Hannah and Paul George, a pair in Kent, England, spent a whole lot of hours attempting to object to an area constructing conversion close to their residence earlier than concluding the system was basically impenetrable with out costly authorized assist. In order that they constructed Objector, an AI instrument that cross-references planning purposes in opposition to coverage to generate formal objection letters in minutes. It permits a person citizen to generate a customized objection package deal in minutes, thereby translating one individual’s real frustration into actionable authorized language.
Besides that native governments are actually bracing for 1000’s of advanced feedback per session. Metropolis planners are legally obligated to learn each single one. When the price of participation drops to close zero, quantity explodes. And each system downstream of that participation—staffed and designed for the previous quantity—experiences course of shock.
Need Radar delivered straight to your inbox? Be a part of us on Substack. Enroll right here.
But when natural participation can overpower these programs, so can manufactured participation. In June 2025, Southern California’s South Coast Air High quality Administration District weighed a rule to part out gas-powered home equipment to chop smog. Board member Nithya Raman urged its passage, noting no different rule would “have as a lot impression on the air that individuals are respiratory.” As a substitute, the board was flooded with over 20,000 opposition emails and voted 7–5 to kill the proposal.
However the outrage was a mirage. An AI-powered advocacy platform referred to as CiviClick had generated the deluge. When the company’s cybersecurity staff contacted a pattern of the supposed senders, they found one thing worrying: Residents confirmed that they had no thought their identities had been getting used to foyer the federal government.
That is the weaponized type of course of shock. The identical infrastructure that lets a Kent couple object to a growth close to their residence additionally lets a coordinated actor flood a system with artificial voices. Confronted with this complexity, the temptation is to easily restore friction. However these previous limitations excluded marginalized contributors. Eradicating them was a real good for society. So the selection isn’t between friction and no friction. It’s between programs designed for people and programs that haven’t but reckoned with machines.
This begins with recognizing that this downside manifests in two basically other ways, every calling for its personal resolution.
The primary is amplification: real customers leveraging AI to scale legitimate issues, flooding the system with quantity, as seen with the Objector instrument. The human sign is actual, there’s simply an excessive amount of of it for any staff of analysts to course of manually. The UK authorities has already began constructing for this. Its Incubator for AI developed a instrument referred to as Seek the advice of that makes use of matter modeling to robotically extract themes from session responses, then classifies every submission in opposition to these themes. As somebody who builds and teaches this know-how, I acknowledge the irony of prescribing AI to remedy the very course of shock it induced. But, a machine-scale downside calls for a machine-scale response. It was trialed final yr with the Scottish authorities as a part of a session on regulating nonsurgical beauty procedures, which confirmed that this know-how works. The query is whether or not governments will undertake it earlier than the subsequent wave of AI-assisted participation buries them.
The second downside is fabrication: unhealthy actors producing artificial participation to fabricate consensus, as CiviClick demonstrated in Southern California. Right here, higher evaluation instruments are inadequate. You can not cluster your solution to fact when the sign itself is counterfeit. This calls for verification. Beneath the Administrative Process Act, federal companies are usually not required to confirm commenters’ identities. That’s the hole the CiviClick marketing campaign exploited. In 2024, the US Home handed the Remark Integrity and Administration Act, which requires human verification to verify that each electronically submitted remark comes from an actual individual. Its sponsor, Consultant Clay Higgins (R-LA), framed it plainly: The invoice’s basis is guaranteeing public enter comes from precise folks, not automated applications.
These are the 2 sides of the identical coin. To successfully deal with this problem, we have to improve the programs that handle public suggestions, whereas additionally strengthening those that confirm its authenticity. Specializing in only one with out addressing the opposite will inevitably result in failure.
Each public system that accepts enter from residents—each remark interval, each zoning assessment, each college board assembly, each insurance coverage enchantment—was constructed on a load-bearing assumption: that one submission represented one individual’s real effort. AI has eliminated that assumption. We are able to redesign these programs to deal with what’s coming, distinguishing actual voices from artificial ones, and upgrading evaluation to maintain tempo with the brand new quantity. Or we are able to depart them as they’re and watch democratic participation change into indistinguishable from AI-generated fakes.

