Your cart is currently empty!


Artificial Intelligence (AI) is rapidly transforming the way we work, research, and interact with the world. From coding applications to searching the web for information, AI promises to revolutionise how we process knowledge. Despite this, there is a glaring flaw in this promise: AI is often surprisingly ill-informed.
The issue? It is struggling to access the most credible and valuable sources of knowledge; peer-reviewed academic journals.
Recently, staff at HERA (the Heavy Engineering Research Association) underwent AI training, exploring the capabilities of AI tools for a variety of tasks. As part of this training, we each developed our own AI agents and apps. Two AI-powered tools that I developed were developed:
The results were both surprising and concerning. The first tool failed to find any copyright breaches, despite known instances existing. More alarmingly, the second tool—designed to find peer-reviewed sources—was struggling to do so. Instead, it returned a mix of cited articles and general website content. This was despite clear and refined instructions specifying that it should focus only on peer-reviewed journal articles, academic books, and conference papers.
At first, this seemed like a technical glitch. However, the real issue became apparent: AI could not access most peer-reviewed research because much of it is locked behind paywalls. This limitation has massive implications for the future reliability of AI.
Academic journals have long served as the gold standard for knowledge verification. Rigorous peer review ensures that published research meets high standards of credibility and accuracy. However, access to this knowledge is often restricted. Publishers charge high fees for journal subscriptions, limiting access to institutions, researchers, and those who can afford it.
AI models, which rely on vast amounts of publicly available data, are effectively shut out from this critical information. Instead, they are left to learn from less credible, freely available sources such as blogs, news articles, and Wikipedia. This creates a major data bias, skewing AI’s understanding of the world.
Consider some practical examples:
The consequences of this data bias extend beyond minor inaccuracies. AI is already being used to inform decisions in many difference fields. If it is relying on incomplete or less credible sources, its recommendations become questionable—at best, incomplete, and at worst, outright wrong.
A stark example is the replication crisis in science, where many published studies fail to be reproduced. If AI models are learning from secondary sources that fail to critically assess these studies, they may propagate false or misleading conclusions. Without direct access to peer-reviewed research, AI is forced to rely on incomplete summaries, biased interpretations, or even misinformation.
The issue raises ethical questions about access to knowledge. In theory, AI has the potential to democratise information, making high-quality knowledge more accessible. However, the current system reinforces a two-tiered structure:
I would suggest that the vast majority of users are in the latter category, with no ability to access primary source documents and make an informed assessment of the reliability (or ludicrousness) of AI-generated knowledge.
This problem is not new. Researchers in developing countries have long struggled with paywalls restricting their access to critical research. The advent of AI now amplifies this issue on a global scale, affecting not just individuals but entire industries and governments that depend on AI-driven insights.
If AI is to truly become a tool for good, we need to address the accessibility of peer-reviewed research. There are a few possible solutions.
Some initiatives, like Plan S (a movement pushing for open-access publishing in publicly funded research), offer hope. Plan S is an initiative by cOAlition S, a group of research funders, to ensure that research funded by public grants is published with immediate open access, aiming to shift the publishing industry towards open access business models. However, widespread change is needed to ensure AI can learn from the best sources, not just the most available ones.
The AI knowledge crisis is not just a technical issue—it’s a societal one. As AI becomes more deeply embedded in decision-making, from business to healthcare, its knowledge gaps could have real-world consequences.
Researchers, policymakers, and publishers need to rethink how knowledge is shared. If AI is to fulfil its promise as an intelligent assistant, it must be trained on the best data available—not just what is freely accessible. Without intervention, we risk an AI future built on biased, incomplete, or even inaccurate knowledge.
It’s time to ask: In the age of AI, who gets to control knowledge? And are we willing to let artificial intelligence be educated by convenience rather than credibility?
It also plays into discussions about research excellence vs research impact. Peer reviewed journals only accept high quality research papers. However, if they are restricting access to that knowledge and quality research, there is certainly an issue of diminished potential research impact. In the age of AI, this is no longer a potential diminished potential for research impact, it is a direct negative impact – as it means AI tools are learning and drawing insights from data that is not necessarily reliable nor based on evidence-based, rigorous science.