The Paradox of Digital Knowledge

Human progress has always depended on shared knowledge. Every breakthrough in science, every cultural shift, every technology that changed daily life was built on what someone else had already figured out. So there is something genuinely strange about the moment we are in. The cost of sharing information is now effectively zero. Sending a research paper to a million people costs the same as sending it to one. And yet access to knowledge has grown more restricted, more fragmented, more conditional. Paywalls, proprietary licenses, DRM: we have rebuilt scarcity in a medium where it no longer exists naturally.

The Economics of Information

Thomas Jefferson got at something important here. When you share a flame, your candle does not go out. When you share an idea, you still have it. Traditional economics is built around the reality of scarcity: if I give you an apple, I am left without one. That logic works fine for apples. It breaks down entirely for information.

Digital technology has pushed this to its logical extreme. Copying a book, a dataset, or a piece of software costs almost nothing. The infrastructure to give every person on Earth access to humanity’s accumulated knowledge already exists. We have simply chosen not to use it that way. Instead, we have layered elaborate mechanisms on top of it: subscription tiers, access fees, artificial restrictions that reimpose the economics of physical objects onto something that has no physical constraints.

The Case for Free Software

Richard Stallman launched the free software movement in the 1980s out of a concrete frustration: he could not fix a printer because the source code was unavailable. The four freedoms the movement codified, running software for any purpose, studying and modifying it, redistributing it, sharing modified versions, were responses to a specific power imbalance. Users had tools they depended on but could not fully understand or control.

I have worked with both proprietary and open systems long enough to feel this asymmetry directly. When software you depend on behaves in unexpected ways and you have no access to the source, you are stuck. You file a support ticket and wait. With free software, you or someone in the community can look, and often fix it. The difference is not theoretical. As software mediates more and more of daily life, from communications to finance to healthcare, the question of who controls it becomes harder to treat as an abstraction. Linux, Apache, Firefox: these are evidence, not just arguments.

Academic Publishing and the Knowledge Barrier

Academic publishing might be the clearest case of how a system can persist long after the logic that built it has dissolved. Researchers, usually funded by public money, produce findings. They submit them to journals, often doing the editorial and peer review work themselves, unpaid. The journals publish the articles and then charge universities, frequently the same institutions that funded the research, for access. A single article can cost more than thirty dollars. Annual institutional subscriptions regularly run into millions.

The consequences are not evenly distributed. Researchers at well-funded institutions can access whatever they need. Those at smaller universities, in developing countries, or working independently often cannot reach basic literature in their own fields. Publishers have positioned themselves as essential intermediaries between authors who want to be read and readers who want to learn, while doing relatively little of the intellectual work. Preprint servers like arXiv have started to crack this open, but they are the exception. The default is still restriction, and that default has real costs.

Cultural Works and the Public Domain

The original logic of copyright was a trade-off: a limited monopoly in exchange for eventual public access. In the United States, the initial term was fourteen years, renewable once. That ceiling has been raised so many times that it now extends to seventy years past the author’s death, longer for corporate works. Each extension has been driven by concentrated interests seeking to hold onto existing assets. There is no evidence that longer terms produce more or better creative work; the research on this is fairly consistent.

What has been lost is harder to measure. Disney built much of its early catalog on public domain stories: Cinderella, Snow White, Pinocchio. Shakespeare adapted plots from earlier writers. Hip-hop developed partly through sampling. Creation feeds on what already exists. When the available material is locked down, the space for new work contracts. The public domain has been essentially frozen since the 1970s, and that is not a small thing.

Digital Libraries and Preservation

The Internet Archive is one of the stranger institutions of our era: a nonprofit in San Francisco that has taken seriously the job of preserving the web, lending digital books, and archiving broadcast media, operating under near-constant legal pressure for doing so. That a project as evidently valuable as this spends so much energy fending off lawsuits reflects something broken in how we have decided to structure knowledge access.

Wikipedia, meanwhile, has become one of the most-visited sites in the world through volunteer labor and open licensing, proving that a different model is viable at scale. Yet vast amounts of knowledge remain beyond what Wikipedia can reach: proprietary databases, out-of-print books whose rights holders cannot be found, academic papers locked behind journal paywalls. Orphan works are a particularly bleak category. These are materials no one is actively selling, whose rights holders are unknown or unreachable, sitting in legal limbo while they deteriorate or remain inaccessible. The law has not caught up with the problem.

The Practical Benefits of Openness

The ethical case for open knowledge gets made often enough. The practical case deserves equal attention. Open systems attract scrutiny, and scrutiny tends to catch problems. The security community has debated “security through obscurity” for decades, and the consensus is clear: more reviewers find more bugs. Open standards also allow diverse implementations, which creates resilience. The internet runs on open protocols that anyone can implement, and that openness is why innovation on top of it has been so explosive.

Proprietary platforms do rise, sometimes dramatically. They also fall, and when they do, they take user data and accumulated content with them. Open knowledge creates dynamics where contributions compound: someone builds on your work, someone builds on theirs, and the whole ecosystem grows. Restriction short-circuits that compounding.

A Path Forward

The honest version of this is: change is possible but the obstacles are real. Copyright extension keeps passing because the industries that benefit from it are organized, funded, and focused, while the public interest in a healthy domain is diffuse and hard to mobilize around. Academic publishers have profit margins that rival the most successful technology companies. They did not build those margins by accident, and they will not surrender them because someone makes a principled argument.

For software, contributing to open source projects matters at the margin, though many engineers have little say over what tools their employers choose. Open source projects often struggle to fund basic maintenance while proprietary alternatives attract significant investment. The pattern is familiar. Linux and Firefox are genuine successes, and a lot of critical infrastructure is still closed.

In academic publishing, mandates from funders have moved the needle. When agencies like the NIH require open access as a condition of funding, journals adapt because they have to. That is a mechanism that works. On copyright reform, the picture is less encouraging. Legislative efforts have mostly failed, fair use is a defense that requires money to assert, and the public domain remains stuck.

I do not write this expecting structural change to follow from individual choices alone. It requires sustained pressure at the political level, which is genuinely hard to organize around issues that feel abstract to most people until they personally hit a paywall at the wrong moment.

Conclusion

The technical capacity to give everyone access to humanity’s accumulated knowledge is not a future possibility. It exists now. The marginal cost is negligible. What stands between that capacity and actual access is a set of political and economic choices, made by identifiable actors serving identifiable interests.

The trajectory over the past few decades has been toward more restriction. Copyright terms have lengthened, academic publishing has consolidated, and DRM has become standard. The open movements have had real wins, but they have not reversed the overall direction. A student in a developing country still cannot access the same materials as one at a wealthy institution, and I see no mechanism that will close that gap soon.

What I keep coming back to is that this is contingent. Nothing about the current arrangement is written into the nature of things. Open source software runs a substantial portion of the internet. Preprints have changed how findings circulate in physics, economics, and biology. Wikipedia, funded by donations and built on volunteer labor, is one of the most-used reference sources in the world. These exist, they work, and they demonstrate that openness is not naive. The question is whether the conditions that enable them can be extended, and that is a political question as much as a technical one.