Can numerical metrics truly measure intellectual value
As modern academia grows increasingly dependent on citations, rankings, and impact factors, the foundations of scientific evaluation risk shifting away from expert judgment, intellectual depth, and meaningful long-term contribution
The modern scientific enterprise stands on a foundation built on trust, intellectual rigour, and collective judgment. Since the early days of journals like Philosophical Transactions (1665), scientific knowledge has been validated through expert scrutiny—an approach that remains central to scientific credibility today.
From Isaac Newton and Robert Boyle to Charles Darwin and James Clerk Maxwell, many of history's most influential scientific contributions were judged not by numerical indicators, but by the informed judgment of fellow scientists.
Even as journals such as Nature and Science emerged, and professional societies institutionalised scientific publishing, the guiding principle remained unchanged: scientific merit must be assessed by experts within the field.
Today, however, this principle is under increasing pressure. Over the past few decades, the evaluation of research and researchers has increasingly moved toward quantitative indicators—impact factors, h-indices, citation counts, journal quartiles, and measures such as CiteScore.
These tools were introduced to manage a rapidly expanding research landscape and to provide a quick, scalable means of assessing research output. For administrators and policymakers, they offer a convenient shorthand—but that convenience comes with inherent limitations.
A closer look reveals the limitations of such metrics. They capture what is easy to measure—citations, publication counts, journal rankings—but often miss what truly matters: originality, methodological rigour, conceptual depth, and long-term impact.
A closer look reveals the limitations of such metrics. They capture what is easy to measure—citations, publication counts, journal rankings—but often miss what truly matters: originality, methodological rigour, conceptual depth, and long-term impact. A paper may accumulate citations for reasons unrelated to quality; a researcher may publish prolifically without making substantive contributions; and, in some cases, journals may inflate their metrics through strategic editorial practices.
Recent cases of journals being reviewed or removed from major indexing databases due to concerns over excessive publication volume, compromised peer review, or failures in quality control highlight the risks of relying too heavily on such indicators. Metrics can provide useful signals, but they cannot substitute for careful judgment.
However, citation-based indicators such as citation counts and h-indices remain meaningful when they genuinely reflect scholarly influence rather than being shaped by strategic practices.
Meanwhile, the transformation of scientific publishing into a global commercial industry has further complicated the landscape. Large publishing houses now manage thousands of journals, generating substantial revenue through subscriptions and author-paid publication charges.
While many maintain strong standards, the scale and speed of operations have introduced new pressures. The rise of open-access publishing, though a welcome step toward wider access, has also brought unintended consequences.
Author-funded models have, in some cases, contributed to the rapid growth of predatory or borderline journals that prioritise revenue over rigorous peer review. Even within established systems, competitive pressures can encourage practices that favour volume over quality.
In such an environment, metrics become both a tool and a target. Researchers—particularly those early in their careers—often operate within systems where numerical indicators strongly influence hiring, promotion, and funding decisions. One visible consequence is the fragmentation of research. Instead of presenting a coherent and comprehensive study, work is sometimes divided into multiple smaller papers to increase publication counts.
While not inherently unethical, this practice can weaken scientific narratives and slow the accumulation of meaningful knowledge. In some cases, it may also contribute to inflated citation patterns—through overlapping content and increased self-citation—thereby artificially enhancing indicators such as citation counts and h-indices.
More broadly, an excessive emphasis on quantity can discourage ambitious, long-term research in favour of safer, incremental studies that are easier to publish. Over time, this may shift the direction of scientific inquiry—from depth and discovery toward productivity and visibility.
Bibliometric comparisons can create an illusion of objectivity. Citation practices vary widely across disciplines; interdisciplinary work is often undervalued; emerging fields lack established citation networks. Even within a single field, metrics cannot reliably distinguish between journals that uphold rigorous standards and those that compromise on quality.
A journal with a high impact factor may not consistently publish robust work, while another with modest metrics may maintain exceptional editorial integrity. Only experienced researchers who are actively engaged in the field can meaningfully assess whether a study advances knowledge and whether its conclusions are sound.
This is why peer judgment remains indispensable. At its best, peer review represents collective intellectual responsibility grounded in expertise, honesty, and critical thinking. While not without limitations, it captures dimensions that metrics cannot: coherence, originality, methodological soundness, and long-term relevance—qualities central to meaningful scientific contribution.
The growing reliance on metrics also reflects deeper structural challenges within academia. Universities, funding bodies, and regulatory agencies must evaluate large numbers of researchers, and metrics offer a convenient and scalable solution.
However, when these indicators come to dominate decision-making, they begin to distort incentives. Researchers may focus more on improving their metrics than on advancing knowledge, while institutions may start to equate productivity with excellence. Over time, this risks rewarding visibility over substance.
This concern is particularly acute in developing countries, where evaluation frameworks often rely heavily on quantitative benchmarks. In such settings, pressure to publish in high-impact journals or accumulate citations can overshadow locally relevant research, practical problem-solving, and capacity building.
In many cases, researchers operate in resource-constrained environments with limited access to external collaborations, where progress depends more on creativity and conceptual insight than on advanced infrastructure. Such contributions are intellectually significant, yet they are often difficult to capture through conventional metrics.
Another important but often overlooked aspect is human resource development. Training students, building laboratories, mentoring young researchers, and creating sustainable research environments are essential for long-term national progress. These efforts strengthen institutions and develop future scientists, yet they are rarely reflected in citation-based indicators. When evaluation systems ignore these contributions, they risk undervaluing what matters most for sustained scientific growth.
The way forward is not to reject metrics, but to use them more wisely. While they can provide useful context, they should not dominate evaluation. Greater emphasis should be placed on expert judgment, focusing on a researcher's most meaningful contributions rather than sheer output. Evaluation systems should also incorporate narrative assessments that consider originality, coherence, and long-term impact.
Institutions must recognise the limitations of such indicators and avoid treating them as definitive measures of quality. Importantly, policy frameworks—especially in developing countries—should acknowledge contributions that metrics fail to capture, including locally relevant research, innovation under constraints, and human capital development. Without such reforms, evaluation systems will continue to reward short-term output over long-term impact.
The history of science shows that knowledge advances through critical thinking, open debate, and intellectual honesty. Metrics, while useful, cannot replace these principles. When over-relied upon, they risk shifting scientific judgment toward algorithms, rankings, and commercially driven benchmarks—undermining the foundations of knowledge itself.
For policymakers, university leaders, and funding agencies, the implication is clear: evaluation systems should place informed human judgment at their core, with metrics used only as supporting tools. This is not merely an academic concern but a strategic imperative. Countries aiming to build strong and innovative research ecosystems cannot depend on narrow indicators that undervalue originality, ignore context, and overlook capacity building.
Ultimately, the issue is not whether to use metrics, but how to use them—and what to value beyond them. Equating numbers with merit risks creating systems that are easy to measure but lacking in meaning. Restoring balance will require institutional courage, thoughtful policy reform, and a sustained commitment to intellectual integrity. Numbers can guide us—but they must not define us.
Dr Mohammed Abdul Basith is a Professor at the Bangladesh University of Engineering and Technology and a researcher in materials science and nanotechnology. He is a Fellow of the Institute of Physics (UK) and the Royal Society of Chemistry (UK), and Founder and Principal Investigator of the Nanotechnology Research Laboratory.
Disclaimer: The views and opinions expressed in this article are those of the authors and do not necessarily reflect the opinions and views of The Business Standard.
