AI Governance Consulting

Systematizing Responsible AI & Governance

Build AI systems your stakeholders, employees, and customers can trust—with governance structures that actually work.

49%

of organizations say they'll institute an AI ethics program

33%

of organizations audit their AI systems for bias

35%

of organizations are prepared for EU regulatory requirements

The Challenge

Why AI Governance Matters

Without clear guardrails, AI systems can introduce bias, security vulnerabilities, legal exposure, and reputational risk. A lack of oversight is one of the leading reasons AI initiatives stall or fail to scale.

AI governance provides the structure needed to responsibly manage AI technologies—defining roles, setting ethical standards, ensuring regulatory compliance, and aligning AI with business goals.

As AI becomes embedded in decision-making from hiring to healthcare, governance ensures systems are fair, secure, transparent, and aligned with human values. With governance, companies can confidently innovate and unlock AI's full enterprise value.

Services

What You Can Do

🏛️

AI Governance Program Development

Adopt a responsible AI governance program that establishes accountability, escalation paths, decision rights, and oversight structures across your AI lifecycle.

⚠️

AI Risk & Impact Assessments

Evaluate risks across your AI use cases using qualitative and quantitative assessments to identify, assess, and mitigate threats while ensuring compliance.

🎓

AI Ethics & Literacy Training

Comprehensive AI ethics and literacy training for employees and stakeholders, enabling them to understand AI's opportunities, risks, and obligations.

🔍

Independent AI Audits

Independent audits to evaluate AI systems for fairness, accuracy, security, and compliance—ensuring accountability and informed governance.

Results

What You'll Achieve

🛡️

Mitigate Risk & Ensure Compliance

Navigate EU AI Act, NIST AI RMF, and internal policies to avoid reputational damage.

📈

Drive AI Governance Maturity

Advance accountability, decision rights, and oversight structures across your AI lifecycle.

👥

Build an AI-Ready Workforce

Create an AI-capable workforce that recognizes opportunities and risks while advancing goals.

INSIGHTS

Trending in Responsible AI

Colorado SB21-169 AI Regulation

Fast Facts: What Professionals Need to Know About Colorado's Insurance SB21-169 Act

May 23, 20252 min read

What is SB21-169?

SB21-169 is Colorado’s groundbreaking insurance regulation requiring insurers to govern their use of external consumer data and algorithms, including artificial intelligence, to prevent unfair discrimination in insurance practices.

Why was it introduced?

To ensure that the increasing use of AI, machine learning, and big data in insurance underwriting, pricing, and claims does not result in bias, unfair discrimination, or harm to consumers—particularly protected classes.

Who does it affect?

All life insurers operating in Colorado who use external consumer data and information sources (ECDIS), algorithms, and predictive models to make decisions about consumers. Other lines of insurance may follow.

How does it work?

  • Requires insurers to establish a governance and risk management framework for ECDIS and algorithms.

  • Insurers must demonstrate that their systems do not result in unfair discrimination.

  • Applies to third-party models and vendor tools used in decision-making.

  • Insurers must submit reports and documentation to Colorado’s Division of Insurance (DOI).

When is it enforced?

Rulemaking was finalized in 2023.

  • Insurers must begin compliance activities and submit their compliance plan in 2024.

  • Enforcement and evaluation of plans begin shortly after submissions.

Where is it applied?

Only in the state of Colorado, but it sets a precedent that other U.S. states may follow, especially as AI regulation gains traction.

Penalties for non-compliance

  • Regulatory action from the Colorado Division of Insurance

  • Potential suspension or revocation of licenses

  • Civil penalties or financial enforcement actions, depending on the violation severity

How does it affect AI systems?

  • AI used in underwriting, pricing, marketing, or claims must be explainable and auditable.

  • Requires documentation of data sources, model design, training, testing, and monitoring.

  • Bias audits and fairness assessments must be conducted.

What should organizations do now to comply?

  • Inventory all models and ECDIS in use—especially those affecting consumer outcomes.

  • Develop a governance framework – include oversight committees, testing protocols, and bias detection methods.

  • Document model development lifecycle – training data, assumptions, limitations, and testing.

  • Conduct bias impact assessments – ensure fairness and non-discrimination.

  • Review contracts with third-party vendors – ensure they meet the Act’s compliance standards.

  • Submit required documentation – align with DOI reporting deadlines and formats.

Finally, Colorado’s SB21-169 is a signal to the insurance industry that AI and algorithmic systems must be fair, transparent, and accountable. Proactive compliance today can position organizations as trustworthy leaders in a rapidly evolving regulatory environment.

For more 5-minute reads that matter, stay tuned for more insights on AI, risk, and governance from Obi Ogbanufe, PhD

AI Regulation
Back to Blog

Let's Talk

Ready to build AI governance at your organization? Let's discuss how I can help you navigate this complex landscape.

Helping professionals build meaningful careers in AI, AI Governance, and organizations build AI systems people can trust.

© 2026 Obi Ogbanufe. All rights reserved.