Enterprise AI Analysis
Building trust in the generative AI era: a systematic review of global regulatory frameworks to combat the risks of mis-, dis-, and mal-information
By Fakhar Abbas, Simon Chesterman, Araz Taeihagh
The rapid evolution of generative artificial intelligence (genAI) technologies such as ChatGPT, DeepSeek, Gemini, and Stable Diffusion offers transformative opportunities while also raising profound ethical, societal, and governance challenges. As these tools become increasingly integrated into digital and social infrastructures, it is vital to understand their potential impact on consumer behavior, trust, information consumption, and societal well-being. Understanding how individuals interact with AI-enhanced content is, in turn, necessary for developing operative regulatory policies to address the growing challenges of mis-, dis-, and mal-information (MDM) on digital platforms. In this study, we systematically analyze global regulatory and policy frameworks as well as AI-driven tools to address the growing risks of MDM on digital platforms and optimize the interplay between humans and genAI moderation. The study highlights the need to balance technological innovation with societal protection and freedom of expression by identifying evolving trends and critical gaps in global policy coherence. We examine how the proliferation of MDM—often accelerated by genAI—distorts the information landscape, induces cognitive biases, and undermines informed decision-making. Our study proposes an integrative strategy that combines technical detection methods with actionable policy recommendations to mitigate MDM risks, reinforce digital resilience, and foster trustworthy genAI governance. The study also explores the potential role of AI itself in combating MDM risks.
Executive Impact Summary
Our analysis quantifies the critical areas demanding attention for robust AI governance.
Deep Analysis & Enterprise Applications
Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.
| Framework | Key Features | Limitations (Cognitive Biases) |
|---|---|---|
| GDPR | Data protection, consumer rights, control over personal data | Doesn't consider cognitive biases impact on data use in disinformation |
| Singapore's Online Safety (POFMA) | Combats online harms (cyberbullying, harassment, false info) | Lacks provisions to counteract cognitive biases in users |
| UK Online Safety Bill | Imposes duties on online platforms to protect users from harmful content | Cognitive biases not considered in its framework |
| EU Digital Services Act (DSA) | Transparency, content moderation, accountability for digital platforms | Doesn't sufficiently address how cognitive biases influence content acceptance |
Challenges and Pathways to Digital Resilience
Impact of AI-weaponized Disinformation in US Presidential Elections
The 2024 US presidential elections experienced significant AI-weaponized disinformation, influencing public opinion. Despite regulatory efforts, enforcement was patchy, allowing harmful content to proliferate. This underscores the need for enhanced international cooperation and stricter penalties for non-compliance, alongside real-time oversight and accountability from tech firms (Deshkar 2024).
Advanced ROI Calculator
Estimate the potential efficiency gains and cost savings by implementing AI-driven MDM mitigation strategies within your organization.
Implementation Roadmap
Our phased roadmap ensures a smooth transition to enhanced digital resilience and trustworthy AI governance.
Phase 1: Regulatory Gap Assessment
Conduct a comprehensive audit of existing regulatory frameworks against genAI-driven MDM risks. Identify critical gaps in standardization, cross-border data governance, and ethical oversight.
Phase 2: AI Tool Integration & Customization
Deploy and customize AI-based detection tools (e.g., watermarking, fingerprinting, GBERT) for identifying synthetic and factually inaccurate content. Focus on multimodal analysis capabilities.
Phase 3: Policy Harmonization & Behavioral Nudges
Develop actionable policy recommendations to mitigate MDM risks, integrating behavioral science insights to address cognitive biases. Promote media literacy and transparency in algorithmic amplification.
Phase 4: International Collaboration & Standards
Engage in multilateral initiatives (e.g., OECD, ISO, UN) to establish common standards for genAI governance and cross-border enforcement, balancing innovation with societal protection.
Phase 5: Continuous Monitoring & Adaptation
Establish real-time monitoring systems for MDM spread and AI misuse. Implement adaptive regulatory frameworks that evolve alongside technological advancements and emerging threats.
Ready to fortify your organization against AI-driven misinformation?
Schedule a personalized consultation with our AI governance experts to develop a tailored strategy for digital resilience.