1. Introduction
Knowledge—of ourselves and the world around us, of laws, customs, duties, and preferences—governs the behavior of individuals and of society. When assessing the current properties directing society’s macro behavioral trends, it is therefore essential to evaluate its common sources of information and knowledge. This also demands consideration of the intricate exchange between them, their methods of production and diffusion, and their forms of reception and maintenance.
The past few decades are commonly referred to as the information age, driven by developments in data transfer and storage capabilities and characterized by a data-based global market in which information has become the greatest available resource [
1]. Luciano Floridi [
2] has coined the terms
infosphere and
inforgs to encapsulate a metaphysical realm of information, data, knowledge, and communication with the former and the informational beings populating it with the latter. Leaving aside the specific component of the digital dimension of knowledge, manipulation of information thus entails dramatic and unprecedented effects on societies globally, the relevance and imminence of which is evidenced by the recent report from the World Economic Forum [
3], which named misinformation the world’s leading short-term threat.
The public relies both on external mediators and internal capacities to navigate the constant influx of information simultaneously arriving from multiple sources and is limited in its ability to assess its quality and reliability. These conditions provide fertile ground for the proliferation of misinformation, much often easier to create, spread, and understand than evidence-based and independently verified data. Sources of misinformation vary from professional literature to social media and commercial sources—each with their respective regulation capacities, incentives, and roles within society. Individual knowledge is therefore the result of a muddle of information pieces with different levels of grounding and credibility.
Problematically, as information becomes more available and accessible, the acquisition of knowledge appears to be diminishing. This information–knowledge gap has significantly widened over time. Misinformation, disinformation, fake news, and other instances of partial, unverified, or simply false information are conceptualized here as a product of this growing gap between knowledge and information, information referring to facts or data that are organized and presented in any meaningful way, and knowledge to a deeper processing and understanding of information. Whereas information usually refers to raw data, does not undergo substantial scrutiny or reflection, and can be false or misleading, knowledge involves more elaborate assessment of information, of which truth is a core property.
Although the precise relationship between knowledge and truth is elaborate and not strictly defined, there is an unequivocal and fundamental bond between these concepts [
4] that is missing from information, not committed to the notion of truth [
5]. Importantly, in the absence of truth or verified knowledge, decision-making relies upon the most readily available sources of information, which are often not those most precise or relevant [
6]. Similarly, decisions made under the condition of uncertainty or stress can often be suboptimal [
7], which can be detrimental in cases involving matters of health and safety. The effects of behavior based on low-quality information manifested in the most direct way during the COVID-19 pandemic [
8]. The impact of misinformation was so dramatic that it was described by some as a separate pandemic related to information with comparable levels of risk termed “infodemic” by the World Health Organization
1.
Who is responsible for keeping individual knowledge in check? As evidence accumulates for the dire consequences of misinformation, national and international bodies attempt to keep up with the rapidly evolving technological landscape and articulate appropriate policies and regulation strategies [
9]. Meanwhile, foreign and private bodies may take advantage of the regulatory vacuum to promote commercial and political interests. Amid the battling powers, individuals are living through these significant changes with little guidance, making it difficult to recognize opportunities and threats or understand the most effective ways to adapt and protect oneself. While the great information tide is lingering, threatening to overtake intellectual freedom and agency, some will inevitably grab onto familiar methods of protection and lean on previously held theories and beliefs.
This paper proposes an epistemic perspective to the harms of information malpractices and borrows the idea of epistemic injustice to demonstrate its effects and suggest possible strategies of engagement. Expanding a term coined by Miranda Fricker [
10], we depict a type of epistemic injustice that transcends these social boundaries. To demonstrate this argument, the current knowledge–information gap is illustrated as giving rise to a new hybrid type of epistemic class—informed ignorance—a position that evolves in an information-rich environment under conditions that do not allow proper acquisition of knowledge.
Clearly, the individual levels and impact of informed ignorance are also shaped by socio-economic and political variables as sources of discrimination and marginalization affecting knowledge accessibility and literacy [
11]. However, although it is relevant to speak about these, we suggest it is not sufficient and position the state of informed ignorance within its cognitive, social, and technological origins and as a core determinant in facing global challenges. In describing its multilayered sources, the epistemic harms are more clearly delineated and understood within their respective fields of origin.
As scientific and technological research and development are supported by public funds, public data, and volunteer participation, which benefit various stakeholders, it is imperative to consistently advocate for the safeguarding of rights to services and data that are optimal, secure, and accessible. This necessitates a fundamental commitment from stakeholders to provide adequately substantiated and easily accessible information concerning the effectiveness, prevalence, and safety of all publicly available and promoted services. While the misinformation crisis is benefiting few but threatening everyone, it is crucial to recognize and engage with all forms of epistemic injustice and oppression. Some international institutions have begun to examine and delineate the main global concerns related to knowledge and misinformation [
9,
12]. In line with these efforts, we attempt to contribute a philosophical and scientific perspective on epistemic injustices, in as much as they impact the global threat of misinformation.
2. The Rise of Informed Ignorance
What are the shared elements that drive informed ignorance and its troubling effects on social order? As the great financial, technological, and political powers steadily push forward, their incentives enhance already encumbering individual cognitive limitations such as anchoring, limited attention, and memory resources [
13,
14] to promote a climate of information overload that heeds the acquisition of knowledge. Other underlying causes of current information gaps and knowledge fallacies are deeply rooted in the modern practice of scientific research [
15]. Exploring the impact of each of these components on shaping individual choices and behaviors is considered here as essential for addressing the risks associated with misguided decision-making based on unfounded information.
2.1. Power, Media, and Access to Knowledge
Undoubtedly, technological development has been one of the main forces shaping and driving the spread of informed ignorance, both by boosting the public availability of information and by catalyzing processes of knowledge democratization. Greater distribution of fiber optic networks, 5G technology, cloud computing, and Content Delivery Networks (CDNs), all contribute to a significantly increased available bandwidth in many regions around the world. While supporting better and faster access to information, such changes also provide continuously evolving ways in which data are collected, distributed, and processed. From this perspective, technology serves as a powerful tool for democratizing information, increasing the availability of information for everyone.
Simultaneously, these technologies allow widespread access to data production and distribution platforms, which may also be employed for the proliferation of misinformation, disinformation, and unverified information and for the selective exposure to information. These can be traced to the concurrent extensive use of recommendation algorithms that utilize personal use data to influence the type of information each user interacts with [
16]. Such algorithms significantly bias individual knowledge formation by deliberately or not reinforcing existing beliefs, limiting doubt and discussion, misrepresenting opposing opinions, and simplifying complex issues. In line with these considerations, advancements in Artificial Intelligence (AI) and deep learning technologies introduce additional problems. Deep fake technologies, for instance, can create highly realistic (yet fabricated) audio and video content, blurring the lines between truth and falsehood.
One of the widely recognized culprits of the current information culture is the global adoption of social media as a prime source of communication and its reciprocal ties with technological development, globalization, and populism [
17]. This foundation gives rise to some of the perpetuators of informed ignorance such as echo chambers, filter bubbles, algorithmic biases and influences, and data manipulation [
18,
19]. Social media algorithms, designed for user engagement, reinforce individuals’ existing beliefs and render them more susceptible to political radicalization [
20]. The rapid dissemination of emotionally charged and sensational content further amplifies these effects, leading individuals toward more extreme political ideologies. Moreover, false or misleading information, often disguised as legitimate news, can distort the comprehension of politically relevant material and allow voters to ‘pick and choose’ views and evidence which align with their ideological inclinations, thus fostering the formation of digital communities that share and reinforce similar ideas, provide a sense of belonging and solidify radicalized political views [
21].
Adding to the complexity and unreliability of information is the intricate and often problematic role played by popular media in the dissemination of information. Incentives driving media sources are often financial and commercial, notoriously lacking in transparency regarding conflicts of interest but also regarding information sources and level of expertise (i.e., in scientific matters). Without such disclosures, media sources further contribute to the spread of unverified information [
22] and disinformation and contribute to a climate of distrust. Importantly, while media reports do not dictate public opinion, numerous studies have demonstrated their substantial influence on the framing of public discourse and the shaping of its attitudes [
23]. Popular media, like social media, by promoting interests besides public awareness and engagement, can introduce bias or oversimplify complex issues. This is especially problematic when mediating complex information to the general public, as in issues concerning healthcare, political debates, and scientific findings.
Furthermore, the sheer speed at which information is disseminated is an important factor in facilitating the dissemination of large volumes of misinformation. As unverified information is much easier to create, and can easily be more sensational, it is more likely to “go viral” and spread before its accuracy is even questioned [
24]. Social media facilitates this phenomenon by rewarding viral content, thus rewarding the creation and spread of more stimulating yet less verified content.
The shift from industrial to technological economic dominance has given rise to new players and powers. With the growing influence of data technologies on decision-making in business, academia, and healthcare, the ability to analyze, navigate, and interpret information has become key in determining individual and company influence and success [
25]. Accordingly, technological companies are becoming increasingly influential in terms of market behavior, while also significantly affecting public discourse and perceptions, through media, communication, and political activity [
26]. On the other hand, the democratization of information, which promotes and supports this technocracy, is maintained by social media, which can also be seen as empowering local and previously marginalized groups and voices [
27].
However, some deem the discrepancies in internet access and digital services both locally and globally as a new form of impairing discrimination, labeled
digital divide [
28]. This term pertains to the uneven distribution of access to internet resources, which in turn has a profound impact on individuals’ opportunities and abilities to engage with information. Importantly, such digital disparity is not isolated but intersects with various social determinants related to technology access such as age, scientific literacy, and political inclination, which play a pivotal role [
29]. More relevantly here, we need to understand how information can be—no matter how counterintuitively—disempowering and, indeed, oppressing.
2.2. Experts, Trust, and Research Integrity
Violations of scientific integrity can promote public skepticism and mistrust by disregarding scientific standards and supporting misinformation. These can be more or less intentional and include methodological and statistical choices, interpretation and publishing biases, hypothesis phishing, and lack of reproducibility [
30]. Some of these practices are deeply rooted within accepted scientific research culture, thus necessitating abundant and mostly unrewarded efforts from researchers in order to completely avoid them. Unfortunately, the cumulative effect of even minor violations over time, combined with less prevalent major violations, results in the publication of unreliable information and an erosion of public trust [
31], the effects of which are far-reaching and untraceable.
Misinformation may also result from common ways of scientific reporting, which involve selective publishing and may favor certain authors or topics based on not-strictly-professional reasoning and publication biases, which may drive researchers to slightly manipulate data to fit a publication or omit less favorable or supportive data. An example from the field of neuroethics relates to the neuroenhancement debate which, despite having produced worthwhile philosophical speculations over its scope and validity [
32], has been shown to conflate or omit certain methods and their contexts of use [
33] in order to promote a prepacked view. Furthermore, for more than a decade, reports have been pointing to discrepancies in published information regarding the prevalence, efficacy, and risks of neuroenhancement use [
34,
35].
Rational ignorance is another way experts can inadvertently manipulate individual choice [
36] by producing information that is not sufficiently accessible because it is either too complicated, too long, or too unengaging. As human beings have limited attention resources that need to be allocated efficiently, it is a harder task to keep one’s attention than to lose it. Hence, stakeholders are required to allocate resources to actively engage users and supply them with the relevant information to make informed decisions. However, this is not in the interest of most stakeholders, who in some cases may even actively or passively benefit from end-user ignorance.
2.3. Cognitive Factors: Limited Resources Meet Excessive Information
An important and seldom addressed contributor to informed ignorance is the function of the human brain or, more specifically, the way in which the human brain processes information, learns, and remembers: the properties of the cognitive system. Human cognition is highly complex and notably involves around 100 billion neurons and an estimated storage capacity of 1.25 × 10
12 bytes, leading to very high processing abilities. Nonetheless, the brain is made of biological matter and is therefore limited in its capacities [
37], including storage size and the speed and accuracy of processing. This idea is captured in the term bounded rationality [
38], which relates such cognitive limits to the ensuing need to rely on less demanding cognitive processes such as biases and heuristics.
Cognitive biases are adaptive functions of data processing but are suboptimal for accurate knowledge production and decision-making. One example is confirmation bias, the tendency to prefer information that confirms previously existing beliefs and reject or avoid contradictory views, traditionally thought to help in avoiding cognitive dissonance [
39]. Another is anchoring, the tendency of individuals to develop over-confidence in an initial viewpoint (e.g., due to a lack of critical skills), which prevents them from updating their beliefs, even in the face of information contradicting their currently held view [
40]. Attention is another cognitive resource that is crucial for information processing. Like others, it is known to be limited in capacity and particularly vulnerable to disruptions and manipulation. In the presence of excessive information, cognitive resources are exhausted, and prioritizing relevant information becomes especially challenging, also due to limited attention capacities [
41], leading again to biased information and misunderstanding.
Such biases can be useful for quickly choosing appropriate behaviors in unfavorable conditions but are not suited for guiding complex learning and analysis of information. Notwithstanding, the current information culture, in which decision-makers often meet conditions of information overload in stressful social and political environments, requires such suboptimal cognitive strategies to be used, exposing individuals to harm and permitting more misinformation to circulate. Reliable information is difficult enough to locate and recognize; however, this becomes more cognitively demanding in the presence of uncertainty, which is a common feature of many knowledge types. This requires users to make an effort to gain relevant knowledge, which is sometimes deliberately made inaccessible, and their ability to do so depends on their level of literacy, technological skills, available time, and so on.
In healthcare, this is easier to observe, as seen in the case of the COVID-19 pandemic, an invaluable example of the power and place of knowledge, information sources, and misinformation. The pandemic created an unprecedented sense of urgency in obtaining and diffusing knowledge under extreme uncertainty, and the effects on public trust and cooperation were immediate [
31]. Group differences in accessibility to knowledge, but also education and cognitive skills, and their effects on perceptions and choices were evident, with dire consequences to social and political stability.
4. Fighting Back
Facing informed ignorance is an urgent task that relies on a collaboration between various fields of knowledge. In recent years, and especially after the COVID-19 pandemic, an important debate is taking place regarding short- and long-term steps required to take in order to contain, thwart, and manage threats associated with information practices. From technological solutions, through awareness campaigns, policy changes, and education initiatives, many approaches have been suggested but few have been tested, such that little evidence regarding their respectable advantages is available [
61,
62,
63]. We wish to add to this pool of resources the call to consider associated epistemic injustices as a core guiding principle.
4.1. Articulating Epistemic Ethics
The unattainability of a solid ground that always favors true knowledge over other kinds of beliefs is fundamental to epistemic thought. But can epistemic theory be utilized to support the war on misinformation? Intuitively, the millennia of epistemic traditionmust have laid the intellectual groundwork for systematically categorizing true and false beliefs. However, this is not so clear in practice. Some philosophical perspectives giving value to knowledge as true belief have been discussed. For instance, Siegmann and Grayot [
64] point to the similarities shared by knowledge leading to true and false beliefs and suggest they are differentiated by the associated understanding involved in acquiring true beliefs.
Importantly, any attempt at articulating an ethics of epistemology must integrate these and similar notions from epistemic theory that address its interaction with normative thought, namely, virtue epistemology. This can be poised within the conditions set forth by the “value turn” in epistemology [
65], traces of which can be found in several of its associated sub-fields. In social epistemology, the role of social processes in the acquisition of knowledge is emphasized and underlies a call for collaborative efforts and diversity of perspectives, while epistemic virtue theory underscores the importance of cultivating intellectual virtues like open-mindedness and humility, which can enhance critical thinking and reduce ignorance. Epistemic notions such as fallibilism and reliabilism also encourage epistemic humility and skepticism by recognizing the limits of human knowledge. Bioethics and medical ethics emphasize the ethical imperative of clear and appropriate communication of relevant information for assuring safe, equitable, and informed decision-making in healthcare. Finally, pragmatic epistemology considers the practical consequences of knowledge, motivating individuals to seek accurate information.
Notably, promoting and voicing the understanding that information is not sufficient for knowledge is fundamental and should be considered within a greater commitment to promote skills for engaging in responsible epistemology. The awareness of the epistemic challenges afforded by current information culture has been referred to as an “epistemic crisis” of post-truth [
66,
67], with suggestions to provide the public with appropriate epistemic tools.
4.2. Practical Steps
Though the practical aspects of facing informed ignorance will most likely lead to gradual and unperceivable changes through long-term effects, they must be articulated regularly and persistently until integrated into appropriate frameworks. The unprecedented challenges encountered during the COVID-19 pandemic to public health systems, global cooperation, clinical research standards, and information dissemination are vital to inform any future strategies. The global infodemic that developed alongside the health crisis afforded a unique opportunity to witness an accelerated version of the looming implications of the bubbling misinformation culture. It is an ethical imperative to consider the indispensable lessons learned from these tragic circumstances in order to promote tangible solutions and avoid future harm, while overlooking this responsibility would amount to the type of epistemic injustice herein discussed.
While the practicalities of the suggested approaches for countering misinformation are not within the scope of this paper and have been extensively elaborated elsewhere, a few major areas of concern will be recounted in line with this paper’s appeal for their consideration within an epistemic framework, namely, media literacy, scientific research standards, well-informed regulation, technological solutions, and stakeholder responsibility.
To provide end-users relevant skills to adapt to the changing information climate, it is necessary to directly engage with public media literacy, while taking into account individual differences and bridging the gap between scientific expertise and public understanding. Promoting literacy can allow individuals to recognize credible sources and be skeptical of unverified information. The necessary conditions for acquiring knowledge, namely, an appropriate psycho-social environment, when extended to the general adult population in the information era must also consider the implications of both external (regulatory, technological, and political) and internal (cognitive training, awareness, and critical thinking) influences. While this is a well-known property in theories of education [
68], it is mostly omitted in the discussions of public discourse.
One important consideration is the incorporation of practices for incentivizing integrity in science and media communication. This involves creating a framework that rewards ethical behavior such as transparency and inclusion. For the scientific community, different strategies have been suggested to promote integrity, from refreshing and reestablishing transparent peer-review processes to improving statistical education and engaging with reproducibility through the scientific reward system. Widely discussed forms of improving scientific standards must be implemented, including open science practices, data sharing, pre-registration of studies, and others, as part of the scientific institutions’ ethical commitment to society.
Addressing informed ignorance through regulation should consider how ethical and intellectual virtues can guide the creation, dissemination, and consumption of information. Regulative actions regarding scientific practices should directly prioritize open and transparent scientific processes in order to encourage a commitment to providing well-founded and evidence-based results while avoiding redundant procedures, whereas regulation regarding data and information management might focus on promoting values of clarity, meaningful organization, and accessible resources. An underlying epistemic virtue shared by regulative practices affecting knowledge production should be a clear commitment and adherence to precision and evidence.
Importantly, social media platforms, as the main arena for public discourse, should prioritize independently verified information and develop creative ways to promote and reward evidence-based information. Some such processes have already been implemented, such as “community notes” in the social media platform “X” (formerly “Twitter”), which is based on community engagement in information confirmation
4 through existing methods of reward, including visibility and popularity. Other systematic procedures such as fact checking and flagging are increasingly utilized but are far from ideal as they are still not immune to exploitation by perpetrators of misinformation.
Finally, stakeholders should be held responsible for promoting and maintaining ethical standards related to information practices, particularly in the context of emerging technologies, media, policymaking, and the actions of scientists, technological companies, and experts. Avoiding the potential dangers leading to informed ignorance is an urgent public safety concern requiring a collective effort from stakeholders across various domains. Influential parties should be continuously scrutinized by external regulation bodies and professional media to detect ways in which they may profit from informed ignorance and may accordingly be short-sighted when determining related governance strategies. As highlighted by Zuboff [
48], identifying the influential players and mechanisms in the new information industry is crucial when outlining the strategies and goals of current societal efforts to protect its values and well-being. Stakeholders, including tech developers, must uphold the commitment to responsible deployment of technology by actively addressing biases, ensuring accountability, promoting fairness in algorithmic decision-making processes, and ultimately by prioritizing ethical considerations over maximizing profits.
5. Conclusions: It Is in Our Hands
Epistemic injustice seems to align with other forms of social and political discrimination; however, beyond this marginalization of certain groups in their ability to impact discourse and knowledge, the idea of epistemic injustice is suggested to extend these boundaries in a meaningful way. The line drawn by power and money, “have and have-nots”, is relevant to understanding some of the forces driving epistemic injustice, but it is not sufficient, and as such may not lead to effective solutions. Perhaps an alternative notion similarly framing the few who hold the means of information production may be more useful in this context. Nonetheless, the processes of knowledge democratization are impacting and shaping the global information market with significant implications.
As the public pushes back against traditional sources of information and expertise, it appears as an act of defiance aimed at taking back control of knowledge production. Accordingly, the new means of information production are free, accessible, equal, and reject accepted structures or rules. However, as in similar historical feats, it is unclear who benefits from these dramatic power shifts. Thus, while still not enjoying the real fruits of labor and common goods associated with the new data-based market dynamics, the public gains only a transient sense of power and freedom. Informed ignorance may be a side effect of the modern counterculture against the intellectual elite, achieving to lower the worth of their work by injecting doubt, mistrust, and skepticism but in the process also losing crucial meaning, nuance, and knowledge.
Informed ignorance, as a human-driven property, calls for actions that are equally within our control. While practical steps are undeniably imperative, they must be underpinned by robust theoretical convictions rooted in clear epistemic virtues. These should unequivocally recognize the social components that drive modern information culture, embrace joint values of truth and justice, which respect social boundaries, and appreciate that our collective understanding is what shapes the trajectory of societal knowledge. By aligning our intentions with these qualities, we can lay the foundation for guidelines and practices that not only undermine the driving forces of informed ignorance but also contribute to the flourishing of a well-informed and enlightened society.