Close Menu
  • Home
  • Alternative News
    • Politics & Policy
    • Independent Journalism
    • Geopolitics & War
    • Economy & Power
    • Investigative Reports
  • Double Speak
    • Media Bias
    • Fact Check & Misinformation
    • Political Spin
    • Propaganda & Narrative
  • Truth or Scare
    • UFO & Extraterrestrial
    • Myth Busting & Debunking
    • Paranormal & Mysteries
    • Conspiracy Theories
  • Contact Us
  • About Us

Subscribe to Updates

Get the latest creative news from FooBar about art, design and business.

What's Hot

Brickbat: Coming Back Clean

April 17, 2026

Twilight of the Strongmen – CounterPunch.org

April 17, 2026

Why the Donald Is Up Shit’s Strait Without a Paddle

April 17, 2026
Facebook X (Twitter) Instagram
Facebook X (Twitter) Instagram
TheOthernews
Subscribe
  • Home
  • Alternative News
    • Politics & Policy
    • Independent Journalism
    • Geopolitics & War
    • Economy & Power
    • Investigative Reports
  • Double Speak
    • Media Bias
    • Fact Check & Misinformation
    • Political Spin
    • Propaganda & Narrative
  • Truth or Scare
    • UFO & Extraterrestrial
    • Myth Busting & Debunking
    • Paranormal & Mysteries
    • Conspiracy Theories
  • Contact Us
  • About Us
TheOthernews
Home»Double Speak»Reframing misinformation as informational-systemic risk in the age of societal volatility
Double Speak

Reframing misinformation as informational-systemic risk in the age of societal volatility

nickBy nickApril 11, 2026No Comments13 Mins Read
Facebook Twitter Pinterest LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Article Metrics

0

CrossRef Citations

When a bank run, a pandemic, or an election spirals out of control, the spark is often informational. In 2023, rumors online helped accelerate the collapse of Silicon Valley Bank. During COVID-19, false claims about vaccines fueled preventable harms by undermining public trust in health guidance, and election lies in the United States fed into the broader dynamics that culminated in the January 6 Capitol attack. These events reveal that misinformation is not just about false or misleading content, but about how degraded information can destabilize entire social systems. To confront this, we must reframe misinformation as an informational-systemic risk that amplifies volatility across politics, health, and security.

By
Nuurrianti Jalli

School of Media and Strategic Communications, Oklahoma State University, USA

IMAGE BY 412designs ON PIXABAY

Misinformation as a systemic threat

Despite growing concern over misinformation’s effects, most institutional responses remain narrow, focused on content removal, media literacy, or identifying “bad actors.” Prior work has examined how misinformation threatens democracy (Tenove, 2020) and how systemic risks cascade across domains (Schweizer, 2021). Yet risk governance frameworks from finance and climate science remain underused for mapping information cascade mechanisms. This commentary bridges that gap, arguing that misinformation should be recognized as an informational-systemic risk, in which degraded or manipulated information flows can destabilize multiple interdependent social, political, and institutional systems, producing effects that cascade beyond the information environment itself.

Systemic risk, a concept from finance and climate governance, refers to disturbances that spread through networks and produce disproportionate failures across a system (Battison et al., 2012; Renn, et al., 2022). Misinformation exhibits similar cascading dynamics, especially when amplified by emotion, opaque algorithms, and speed. Recent episodes make this clear: viral rumors influenced the rapid 2023 collapse of Silicon Valley Bank (Khan et al., 2024), and election misinformation has fueled institutional delegitimization and unrest, from the January 6 attack in Washington (Wang, 2022) to post-election violence in Jakarta in 2019 (Temby, 2022). These examples suggest that feedback loops are recurrent, observable, and potentially destabilizing.

Rather than appearing as isolated incidents, these dynamics reveal how misinformation interacts with fragile systems. Misinformation exploits political distrust, financial volatility, and public health crises, turning points of vulnerability into wider disruption. This is what makes misinformation a systemic risk: it does not remain contained within one domain, but can cascade across sectors, triggering instability in ways that are increasingly difficult to anticipate or control.

Today, we are entering a period of mounting informational fragility. Trust in institutions is declining globally (Valgarðsson et al., 2025), and platform governance remains unevenly regulated across contexts, ranging from structured regulatory frameworks such as the European Union’s (EU) Digital Services Act to environments shaped by heavy-handed regulation or censorship (Jalli, 2024). Against this backdrop of declining institutional trust and uneven platform governance, AI-generated content, from deepfakes to synthetic propaganda, circulates widely at low cost, weakening the system’s resilience. Resilience refers to the capacity of information systems, including platforms, institutions, and publics, to withstand shocks without experiencing critical failures. In practice, this may involve stress-testing algorithms for vulnerability to viral falsehoods, establishing rapid response coordination centers within governments, and investing in trusted communication channels across civil society. Addressing this challenge requires more than correcting individual falsehoods—it demands strengthening the resilience of information systems.

From episodic error to systemic fragility

Traditional approaches tend to treat misinformation as an isolated error rooted in bias, poor judgment, or manipulation. These episodic framings, while useful, struggle to explain how misinformation spreads in environments where institutional, technological, and social systems are deeply interdependent. Systemic risk analysis helps capture these dynamics by showing how disruptions propagate through networks and interact with hidden vulnerabilities. Viewed through this lens, misinformation is not simply misleading content but a structural force that can weaken the trust and legitimacy on which societies depend. The defining feature of systemic risk is not its scale but relational dynamics, particularly its ability to spread through networks, interact with latent vulnerabilities, and push stable systems toward crises (Renn et al., 2022; Schweizer, 2021). Misinformation fits this model. Its danger lies less in single falsehoods than in its capacity to undermine the informational foundations that sustain democratic function, including confidence in science, the legitimacy of elections, and shared historical narratives. Extending systemic risk analysis to the informational domain helps clarify why misinformation should be understood as a structural condition rather than an episodic anomaly.

These dynamics are visible across domains. In 2017, false narratives and hate speech circulated on Facebook helped justify attacks against the Rohingya minority in Myanmar, showing how online misinformation can escalate into real-world atrocities (Amnesty International, 2022). In the years that followed, Myanmar authorities imposed internet shutdowns and social media restrictions, citing security concerns but in practice restricting freedom of expression (Human Rights Watch, 2019).

Figure 1. Myanmar information crisis and systemic cascade (2016–present). Platform moderation failure escalated to mass atrocities, triggering platform bans that prompted state retaliation, blocking civilian access. Ongoing shutdowns (130+ in 2024) illustrate cascading failures and feedback loops characteristic of systemic risk.

This sequence suggests that responses to misinformation crises can themselves act as catalysts for systemic vulnerability. Facebook’s ban of military accounts prompted junta retaliation that blocked civilians’ primary communication tool while the military continued to exploit the platform for propaganda (Freedom House, 2024). Misinformation can also disrupt disaster responses and climate shocks. During Hurricanes Harvey and Irma, false rumors about shelter policies and mandatory ID checks spread widely on social media, discouraging vulnerable groups from seeking safety and forcing officials to divert resources to debunking misinformation (Hunt, Wang, & Zhuang, 2020). During Australia’s 2019-2020 bushfires, coordinated online narratives exaggerated the role of arson, amplified through the hashtag #ArsonEmergency, distracting attention from climate drivers and weakening support for mitigation measures (Weber et al., 2020). These examples show how misinformation can cascade across environmental hazards, complicating emergency management and climate governance, and contributing to the escalation of natural risks into systemic crises. This commentary extends that pattern of analysis by applying systemic risk governance frameworks from finance and climate studies to examine how misinformation generates cascading failures across domains and should be treated as a cross-infrastructure vulnerability rather than an isolated content problem.

Figure 2. Misinformation as a self-reinforcing systemic risk. Misinformation functions not as a one-off disruption but as a recursive, systemic threat. It undermines fragile information infrastructures, destabilizes institutions, fractures social trust, and weakens governance. These breakdowns, in turn, amplify the production and circulation of misinformation, creating feedback loops that demand structural, not episodic responses.

To understand why misinformation tends to cascade across domains, it is necessary to look beyond outcomes and examine the infrastructures that generate them. The sociotechnical perspective shows how platform architectures and design incentives make misinformation not an anomaly but a predictable feature of today’s information environment. Here, infrastructures refer to the shared technical and organizational substrates, such as platforms, networks, and protocols, that enable diverse local systems to operate. Systems, on the other hand, are the locally bounded processes and institutions that function within and through those infrastructures with specific goals and internal interactions (Edwards, 2013).

Misinformation as a sociotechnical threat

Misinformation arises not only from individual behavior or malign actors but also from the infrastructures that govern digital communication. Platform research finds that algorithmic design and advertising logic systemically privilege attention-grabbing content, regardless of accuracy (Gillespie, 2018; Vaidhyanathan, 2018). Recommendation systems optimized for engagement can amplify emotionally charged posts, while monetization models reward divisive content that keeps users active. Far from neutral, these systems normalize the circulation of misleading content by embedding it into everyday communication flows.

Studies of YouTube, Facebook, and Twitter indicate that attention-driven architectures promote emotionally charged and polarizing material, often elevating misinformation above verified information (Tufekci, 2015; Vosoughi et al., 2018). Such dynamics make misinformation a predictable outcome of sociotechnical design rather than an anomaly.

As these architectures evolve, the rapid integration of generative AI into content production and distribution further increases existing risks. Synthetic text, images, and video, produced at scale and often indistinguishable from authentic material, can intensify misinformation campaigns (Paris & Donovan, 2019; Weidinger et al., 2021). Detection systems often lag behind innovation, while moderation practices remain fragmented and uneven across platforms and can leave exploitable gaps. The concern is not only that people may believe false claims, but that digital systems themselves are structured in ways that enable, incentivize, and monetize them. Effective governance must therefore address both the technical architectures that accelerate misinformation and the broader social conditions that allow it to persist.

Deepfakes, AI, and the informational arms race

Deepfakes illustrate how generative AI is compounding systemic risk in the information environment. Unlike earlier manipulated media, deepfakes can generate audio and video that appear indistinguishable from authentic recordings, creating major challenges for verification (Paris & Donovan, 2019). These threats to evidentiary trust affect not only individuals but also courts, newsrooms, and democratic institutions that depend on credible records. By complicating verification processes, deepfakes can amplify the cascading dynamics of systemic risk, potentially weakening institutional legitimacy, undermining governance, and heightening geopolitical insecurity.

These tools are appearing more frequently in strategic information operations. During Taiwan’s 2024 presidential election, a fabricated audio clip alleging that a senior politician accused Lai Ching-te of embezzlement was confirmed as a deepfake. The same campaign used spliced and overdubbed video clips to distort public statements. Reportedly linked to state-aligned media networks, these tactics show how synthetic media is being adapted for election interference, straining fact-checking systems and eroding trust in democratic processes (Hung et al., 2024).

The United States has faced similar risks. In May 2025, the FBI warned that AI-generated voice messages and texts were being used to impersonate senior officials in attempts to harvest credentials and compromise networks (Vicens, 2025). Such incidents suggest that deepfakes are evolving beyond curiosities or isolated hoaxes and are increasingly integrated into the infrastructure of political disruption. Their spread challenges the traditional notions of proof, undermines verification norms, and deepens uncertainty in everyday discourse. Deepfakes, therefore, may transform misinformation from isolated episodes of deception into a bigger systemic risk. By threatening the evidentiary foundations on which institutions and publics rely, synthetic media can accelerate cascades across domains, turning fragile systems in health, finance, and politics into contested arenas over reality.

The limits of behavioral fixes

In response to the global misinformation crisis, education and media literacy remain the most common solutions. The assumption is that if people are taught to spot falsehoods, they will be more resilient to manipulation. This approach is appealing because it is scalable, empowering, and politically uncontroversial. Yet in today’s fragmented and fast-moving environment, its limits are increasingly evident.

Misinformation is not merely about what people believe, but also about the systems that shape those beliefs. Even highly informed individuals are continually exposed to misleading content, some algorithmically prioritized and some synthetically generated, much of it circulating through attention-driven economies (Guess, et al., 2020; Pennycook & Rand, 2021). Knowing how to evaluate sources cannot stop an AI-generated video from going viral or an emotionally charged conspiracy from dominating timelines. The challenge lies less in individual capacity than in the structural conditions that overwhelm it (Guess & Lyons, 2020).

Many interventions rely on a “deficit model” that assumes people fall for misinformation because they do not know better. Yet research shows that belief in falsehoods is often driven by identity, emotion, and belonging rather than ignorance (Bail, 2022; Lewandowsky et al., 2017; Kahan, 2017). People share false claims not simply because they are misled, but because those claims feel true, reinforce loyalties, or signal political alignment. No fact check can easily compete with that kind of resonance, especially when reinforced by algorithmic amplification and social feedback (Nyhan & Reifler, 2010).

The tension between social motivations for sharing and systemic risk is crucial. Systemic risk does not necessarily require widespread belief in false claims; it can emerge through multiple pathways. Widespread circulation (regardless of individual belief) tends to degrade information quality and increase the cost of discerning truth (Vosoughi et al., 2018). Even limited belief among key actors like policymakers, journalists, or health officials may trigger cascading failures. Endemic uncertainty can constrain institutional decision-making, while resources diverted to fact-checking and counter-messaging often weaken institutional capacity (Helbing, 2013). Social sharing thus can amplify rather than mitigate systemic risk by accelerating distribution and saturating information networks, making misinformation a structural vulnerability that extends beyond individual credulity.

Behavioral interventions remain valuable but cannot serve as the primary response. They are too reactive, slow, and overly focused on individuals while neglecting the infrastructures that reward virality over accuracy. Media literacy programs do not change the fact that content moderation remains inconsistent, underfunded, and politicized. If misinformation is a system-wide vulnerability, responses should include regulatory frameworks that address design incentives, algorithmic transparency, and platform accountability. Information integrity should be treated as a public good rather than an individual responsibility.

Reframing governance for systemic resilience

If misinformation is embedded in sociotechnical systems rather than isolated content, governance responses must evolve accordingly. The challenge is not only what people post, or believe, but how digital infrastructures are designed, governed, and incentivized (Chesterman, 2025). Systemic risks demand systemic thinking, yet current strategies remain misaligned with the scale and complexity of the threat (Helbing, 2013).

Governance today is fragmented and often reactive. Content moderation struggles to keep pace with the volume and velocity of falsehoods. Laws targeting misinformation are frequently overly broad, vulnerable to political misuse, or narrowly focused on criminal intent. International coordination remains limited, while the platforms that shape global discourse operate across borders with uneven transparency and accountability (Gorwa, 2019). This mismatch creates conditions under which misinformation flourishes not only because it is persuasive but because the environment rewards its circulation.

Enhancing resilience begins by recognizing platforms as critical infrastructure rather than neutral conduits. Algorithmic design, moderation, and enforcement policies influence democratic processes, public health, and geopolitical stability. Operationally, platforms could be required to conduct adversarial testing that simulates how false narratives about elections or health might spread through recommendation systems, identifies amplification points, and demonstrates mitigation capacity, similar to financial stress tests. Although some platforms conduct internal testing, transparency and enforcement remain limited (Urman & Makhortykh, 2023).

The EU’s Digital Services Act (DSA) requires large online platforms to conduct annual systemic risk assessments and independent audits of their content-governance practices (European Commission, 2022a). While the companion Digital Markets Act (DMA) addresses competition and gatekeeper power, comparable obligations remain rare outside the EU (European Commission, 2022b). Effective implementation depends on international coordination, as platforms operate across jurisdictions and unilateral regulations risk regulatory gaps.

Just as governments regulate financial systems to prevent cascading economic failure, there is a case for meaningful oversight of how information is distributed, ranked, and monetized (Gillespie et al., 2020). Building systemic resilience, therefore, requires infrastructural reforms, platform accountability, and cross-border cooperation. Without stronger coordination, interventions will remain reactive, leaving societies vulnerable to future informational shocks.



Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
nick
  • Website

Related Posts

How Russia weaponizes the church in Ukraine

April 14, 2026

Not so different after all? Antecedents of believing in misinformation and conspiracy theories on COVID-19

April 11, 2026

People are more susceptible to misinformation with realistic AI-synthesized images that provide strong evidence to headlines

April 11, 2026
Leave A Reply Cancel Reply

Demo
Our Picks

Putin Says Western Sanctions are Akin to Declaration of War

January 9, 2020

Investors Jump into Commodities While Keeping Eye on Recession Risk

January 8, 2020

Marquez Explains Lack of Confidence During Qatar GP Race

January 7, 2020

There’s No Bigger Prospect in World Football Than Pedri

January 6, 2020
Stay In Touch
  • Facebook
  • Twitter
  • Pinterest
  • Instagram
  • YouTube
  • Vimeo
Don't Miss

Brickbat: Coming Back Clean

Politics & Policy April 17, 2026

A local TV news investigation found that hundreds of people in Georgia who were arrested for…

Twilight of the Strongmen – CounterPunch.org

April 17, 2026

Why the Donald Is Up Shit’s Strait Without a Paddle

April 17, 2026

The Political Culture of the Smartphone and the Cult of the Algorithm

April 17, 2026

Subscribe to Updates

Get the latest creative news from SmartMag about art & design.

Facebook X (Twitter) Instagram Pinterest
© 2026 ThemeSphere. Designed by ThemeSphere.

Type above and press Enter to search. Press Esc to cancel.