Importance of AI governance: A critical imperative for Sri Lanka

Thursday, 26 February 2026 04:00 -     - {{hitsCtrl.values.hits}}


Introduction: Why AI governance matters now

Artificial Intelligence (AI) is no longer a futuristic concept—it is already embedded in decision-making across finance, healthcare, logistics, education, public administration, and national security. While AI offers immense productivity gains, efficiency, and innovation, it also introduces significant ethical, legal, operational, and security risks.

As AI adoption accelerates globally, AI governance has emerged as a critical mechanism to ensure that AI systems are safe, ethical, transparent, and accountable. Countries that fail to establish robust AI governance frameworks risk data breaches, reputational damage, regulatory non-compliance, and erosion of public trust.

For Sri Lanka, which is actively pursuing digital transformation and e-governance initiatives, the absence of a structured AI governance framework could become a strategic vulnerability rather than a development advantage.

Global AI governance landscape: Lessons for Sri Lanka

Many countries have already recognised the urgency of governing AI responsibly and have introduced national or regional frameworks.

Established International AI Governance Frameworks

Australian AI Ethics Framework

Focuses on principles such as fairness, transparency, accountability, privacy protection, and human-centred AI.

European Union AI Act (2024)

Introduces a risk-based approach, classifying AI systems into unacceptable, high-risk, limited-risk, and minimal-risk categories, with strict compliance obligations for high-risk applications. The AI Act is a European regulation on artificial intelligence (AI) – the first comprehensive regulation on AI by a major regulator anywhere.

NIST AI Risk Management Framework (USA)

Provides a voluntary but structured framework to identify, assess, manage, and govern AI-related risks across the AI lifecycle.

China AI Governance Framework

Strongly emphasises state control, algorithm accountability, content moderation, and alignment with national security priorities.

Singapore Model AI Governance Framework

Widely regarded as a practical, business-friendly framework focusing on explainability, transparency, human oversight, and responsible deployment.

 ISO/IEC AI Standards

Including ISO/IEC 23894 (AI risk management) and ISO/IEC 42001 (AI management systems), which provide globally recognised governance and control structures.

Key takeaway:

AI governance is no longer optional—it is becoming a regulatory expectation and competitive differentiator.

Key risks associated with AI usage

While AI improves efficiency, unmanaged AI usage exposes organisations to serious risks:

1. Data confidentiality and security risks

One of the most critical risks arises from employees uploading sensitive information into public AI platforms such as ChatGPT.

Real-World Incident: US Government Data Exposure (2026)

In early 2026, a senior official at the Cybersecurity and Infrastructure Security Agency (CISA) uploaded “For Official Use Only” government contracting documents into a public version of ChatGPT for policy review purposes.

The incident triggered automated internal security alerts

The uploaded data was sensitive but unclassified

It represented a direct breach of internal security protocols

The official involved was acting CISA official Madhu Gottumukkala

This incident highlights how well-intentioned productivity use of AI can result in severe governance failures.

2. Data exposure and model training risks

When data is submitted to public AI models:

The data may be stored

It may be used to train future models

Sensitive information could resurface indirectly in future AI-generated responses

Even if identifiers are removed, contextual or proprietary knowledge leakage remains a significant concern.

3. Widespread organisational risk

Research by Cyberhaven indicates: 8.6% of employees have pasted company data into ChatGPT, with 11% of all data pasted being classified as confidential. More concerning, 4.7% of employees have pasted sensitive data, including source code, client information, and strategic documents. (Cyberhaven, 2025)

This demonstrates that AI-related data risk is systemic, not accidental or isolated.

The case for organisational-level AI governance policies

To mitigate these risks, organisations must move beyond ad-hoc controls and establish formal AI governance policies.

Essential organisational AI controls

Clear policies defining what data can and cannot be uploaded to AI tools

 Mandatory data classification awareness

Approval processes for AI usage in sensitive functions

Employee training on AI ethics and data responsibility

Clear accountability for AI-related breaches

Secure AI alternatives and mitigation measures

While individual users can:

Disable chat history- Users can turn off chat history so their conversations are not saved or used to train the AI.

opt out of data training- Users can choose not to allow their inputs to be used to improve future AI models, helping to reduce data reuse.

These measures are insufficient at enterprise or government level.

Many organisations are therefore adopting:

Microsoft Azure OpenAI Service with enterprise agreements

Private or on-premise AI models

Contractual guarantees on data isolation and non-training

These solutions allow organisations to leverage AI benefits without compromising confidentiality.

Why Sri Lanka urgently needs a National AI governance framework

Sri Lanka’s increasing use of AI in:

Banking and financial services

Public sector digitalisation

Education and professional training

Logistics, healthcare, and telecom

…makes the absence of a national AI governance framework a significant policy gap.

Key risks for Sri Lanka without AI governance

Data privacy violations (especially under the Personal Data Protection Act)

Unethical or biased AI decision-making

Loss of public trust in digital government initiatives

Reputational damage in international partnerships

Regulatory misalignment with global markets

Recommended AI governance directions for Sri Lanka

Sri Lanka should consider:

1. A National AI Governance Framework aligned with:

  •  EU AI Act principles (risk-based regulation)
  • NIST risk management concepts
  • ISO AI management standards

2. Sector-Specific AI Guidelines

  • Banking and finance
  • Government and defence
  • Healthcare and education

3. Mandatory AI Usage Policies for public institutions

4. Enterprise-Level AI Governance Requirements

  • AI risk assessments
  • Data governance integration
  • Audit and accountability mechanisms

5. Capacity Building

  • Training policymakers, regulators, and organisational leaders on AI risks and governance

Conclusion: Governing AI to unlock its true value

AI governance is not about restricting innovation—it is about enabling sustainable, trusted, and responsible AI adoption.

The 2026 CISA incident demonstrates that even advanced economies are vulnerable without strong AI governance. For Sri Lanka, proactive action today can prevent costly failures tomorrow.

By learning from global best practices and tailoring them to local realities, Sri Lanka can position itself as a responsible AI adopter, safeguarding data, protecting citizens, and enhancing economic competitiveness in the AI-driven world.


(The author is a Senior Chartered Accountant with over 20 years of professional experience, primarily in the banking sector, and is currently serving as AGM– Audit at one of the country’s leading banks. He is also a CISA-certified auditor (ISACA), holds professional qualifications in Artificial Intelligence, and serves as a visiting lecturer at PIM, CA Sri Lanka, and IBSL) 


 

Recent columns

COMMENTS