How AI is Poisoning the Free and Open Source Software Movement
Indrajeet Patil
FOSS developer, 8+ years

Photo: Daniel Prado / Unsplash
Source code for these slides can be found on GitHub.
96%
of codebases include open-source software.
Harvard summary of the 2024 working paper
$8.8T
estimated global replacement value if firms had to recreate the OSS they rely on.
Harvard demand-side estimate
2-5x
reported ROI for organisations that actively contribute upstream instead of only consuming OSS.
Linux Foundation, 2026
Modern software runs on FOSS. Pollute the commons, pay a systemic cost.
Review labour
Volunteer maintainers turn raw patches into shared infrastructure.
Trust
Contributors usually begin with a presumption of good faith.
Apprenticeship
Small PRs are how future maintainers usually enter the project.
LLMs change the cost model of all three at once.
Note on pre-AI FOSS: It is important not to romanticize pre-AI FOSS. Low-quality PRs, drive-by contributors who disappear after dumping code, and maintainer burnout existed long before LLMs. AI just makes the firehose of low-quality contribution cheaper and more plausible-looking. The root issue is still the tragedy of the commons + under-compensation of maintainers, not AI per se.
+5.9%
Copilot use was associated with higher OSS code contributions, even as coordination time also rose.
Triage relief
Andrea Griffiths: “AI can make maintainership sustainable” when it takes a first pass at repetitive issue triage.
Good AI use exists
Ghostty accepted AI-assisted reports that transparently explained the process and helped fix four real crashes.
This talk is not anti-AI. It is about what happens when speed, anonymity, and volume outrun stewardship.
Who captures the value, and who carries the cost?
The paradox is fake
FOSS still needs more human contributors. What it cannot absorb is more anonymous, low-context synthetic volume.
What healthy contribution used to create
Not all teams absorb this equally
React can lean on engineers paid by Meta. Independent maintainers and small volunteer teams usually have no such buffer.

The old loop
January 2026
Adoption rose. The docs-led business model cracked.

Sources: Adam Wathan on PR #2388 · Tailwind Sponsor · Tailwind Plus
Causation vs. correlation: Tailwind’s traffic drop could also stem from better general-purpose AI search, market saturation, or competitors. AI isn’t necessarily the sole driver. Similarly, “extraction without reciprocity” isn’t purely one-way; many AI labs open-source their models and tools.
Example: Malus
An AI “clean room” service claims it can recreate open-source packages from docs and APIs.
The pitch
Keep equivalent functionality, but remove attribution, copyleft, and license obligations.
The liability is no longer only bugs in public code. It is the risk that openness becomes a reconstruction guide.
The fear is not that open source suddenly stops working. It is that AI lowers the cost of routing around the obligations that made sharing sustainable.
Attack surface, supply chain, and code provenance.
Noise became part of the incident.
March 24, 2026
LiteLLM issue #24512 warned that the package on PyPI was compromised.
488 comments
The thread was submerged in repetitive bot-shaped replies.
12:49 UTC closed as NOT_PLANNED.
13:49 UTC reopened.

Sources: BerriAI/litellm issue #24512 · issue #24518
Compression
Recon, exploit drafting, follow-up comments, and social engineering now collapse into minutes.
Parallel work
One attacker can write code, write prose, and flood response channels at the same time.
Synthetic volume becomes an attacker advantage.

Photo: Maxim Tolchinskiy / Unsplash
Plausible is not traceable.
Generated code can look mergeable while still hiding its origin, licence lineage, or attribution trail.
The next step is worse.
Malus is a satirical warning, but it captures the logic: AI clean-room claims can be used to dodge attribution and copyleft obligations.
Maintainers answer with attestation, provenance checks, or outright bans on AI-derived code.

The package does not need to exist first.
The model can invent it. An attacker can register it later.
5.2%
average rate for commercial coding models.
21.7%
average rate for open-weight coding models.

Sources: USENIX ;login: (Jun. 17, 2025) · USENIX Security ’25 · 205,474 phantom names observed
Review queues, trust, and the contributor pipeline.
Cheap to submit.
Expensive to review.
Public example, February 2026
Matplotlib maintainer Scott Shambaugh rejected an AI-authored change. Hours later, an AI persona published a personal attack piece about him.

XZ Utils changed the baseline
The lesson was not “review newcomers less kindly.” It was that trust itself can be weaponized.
AI flood changes the front door
When maintainers see more bot-shaped PRs, every unsolicited submission becomes costlier to parse and easier to distrust.
This is why the “we want contributors, but reject many PRs” tension is real only on the surface: projects still want people, but they can no longer afford unaccountable volume.
Mitchell Hashimoto is piloting a different model.

Ghostty: first-time contributors need a maintainer vouch before a PR can stay open.

Vouch: trust becomes explicit before code arrives.
Sources: Ghostty CONTRIBUTING.md · mitchellh/vouch
tldraw, January 2026
An influential project announced it would automatically close pull requests from external contributors.
The repo stayed open for issues, bug reports, and discussions. The part that narrowed first was the expensive part: reviewable code contribution.
This is what “the commons hardening” looks like in practice: not total closure, but higher gates at exactly the points where AI volume is most costly.
GitHub’s own availability post frames the problem as a platform-scale shock: agentic development turns every pull request into load on storage, CI, search, permissions, notifications, and APIs.
Junior developers
Lose first reviews, apprenticeship, and the easiest route into open-source work.
Maintainers
Carry more suspicion, more context switching, and more unpaid incident response.
Projects
Lose diverse small contributions and become more closed and fragile.
The movement
Loses the renewal mechanism that kept FOSS broad and resilient.
Technical Counter-measures
Social & Policy Shifts
The open-source ecosystem is actively adapting. Emerging solutions suggest the current disruption will eventually stabilize into a new equilibrium.
Protect review time
Protect provenance
Protect the commons
If openness becomes unaffordable, FOSS stops regenerating.
Additional references: XZ Utils backdoor summary · Mitchell Hashimoto on AI and open source
Questions and critiques welcome.
Source code for these slides is available on GitHub.
See more slide decks on software engineering and open source.