Back to Blog
AI Governance AI Policy Shadow AI Risk Management AI Maturity Business Policy Responsible AI

The Human-in-the-Loop: A Governance Framework for Aussie SMBs

By Ash Ganda | 17 January 2026 | 11 min read

Your employees are using AI right now. The question isn’t whether—it’s whether you know about it, control it, or are exposed to risks you haven’t considered.

“Shadow AI” is the term for unauthorised AI tool usage in organisations. It’s the receptionist pasting customer complaints into ChatGPT. It’s the sales rep using an AI email writer with your CRM data. It’s the accountant uploading financials to an AI analysis tool.

Each instance is potentially a privacy breach, a security hole, or a compliance violation. And it’s happening in almost every Australian business.

At CloudGeeks, we’ve helped businesses transform AI chaos into AI governance. Here’s how to create a “Human-in-the-Loop” framework that harnesses AI benefits while managing risks.

The Shadow AI Problem

What’s Actually Happening

Research from 2024 indicates that over 60% of employees have used generative AI for work purposes. Most did so without employer knowledge or approval.

Common Shadow AI Activities:

  • Writing or editing emails and documents
  • Summarising meeting notes
  • Analysing customer feedback
  • Generating marketing content
  • Coding and technical work
  • Creating presentations
  • Financial analysis and modelling

Why It’s Dangerous

Privacy Breaches: Personal customer data pasted into external AI tools may violate privacy laws.

Data Leakage: Sensitive business information sent to public AI tools becomes potential training data.

Accuracy Risks: AI outputs used without verification can propagate errors into business decisions.

Legal Liability: AI-generated content may infringe copyright or contain defamatory statements.

Contract Violations: Using AI may breach client confidentiality agreements.

Regulatory Exposure: Certain industries have specific requirements around AI usage.

Why Employees Do It

Don’t assume malice. Employees use Shadow AI because:

  • It genuinely makes them more productive
  • They don’t know it’s problematic
  • No approved alternatives exist
  • They face pressure to deliver more, faster
  • They’ve seen others do it without consequences

Understanding motivation helps design policies that work.

Assessing Your AI Maturity

Illustration showing Shadow AI risks with employees using unauthorized AI tools like ChatGPT to process customer data, financial information and confidential business documents creating privacy breaches, data leakage and compliance violations for Australian businesses

Before creating governance, understand where you stand. The Deloitte AI Maturity Model provides a useful framework.

The Five Levels

Level 1: AI Unaware

  • No formal AI strategy
  • Shadow AI likely prevalent
  • No policies or guidelines
  • Reactive approach to AI risks

Level 2: AI Aware

  • Leadership recognises AI importance
  • Initial exploration of AI tools
  • Basic awareness of risks
  • Beginning policy development

Level 3: AI Activated

  • Specific AI use cases implemented
  • Formal AI policies in place
  • Designated AI responsibility
  • Measured outcomes

Level 4: AI Operationalised

  • AI integrated into core processes
  • Comprehensive governance framework
  • Regular monitoring and adjustment
  • Clear ROI demonstrated

Level 5: AI Transformational

  • AI embedded in strategy
  • Continuous innovation with AI
  • Industry-leading practices
  • AI as competitive advantage

Self-Assessment Questions

Score yourself 1-5 on each dimension:

Strategy

  • Do you have a documented AI strategy?
  • Is AI discussed at leadership level?
  • Are AI investments prioritised?

Data

  • Is your data AI-ready (clean, accessible)?
  • Do you know where sensitive data is?
  • Can you audit data usage?

Technology

  • Have you evaluated AI platforms?
  • Do you have approved AI tools?
  • Can you enforce technology policies?

People

  • Do staff understand AI capabilities and risks?
  • Is there designated AI responsibility?
  • Can employees get AI guidance?

Governance

  • Do written AI policies exist?
  • Are there consequences for violations?
  • Is AI usage monitored?

Score 5-10: Level 1 (Unaware) Score 11-15: Level 2 (Aware) Score 16-20: Level 3 (Activated) Score 21-25: Level 4 (Operationalised)

Most Australian SMBs score Level 1-2. Don’t be discouraged—awareness is the first step.

Assessing your AI Maturity using the Deloitte framework

Creating Your AI Acceptable Use Policy

Policy Structure

A practical AI Acceptable Use Policy includes:

  1. Purpose and Scope
  2. Definitions
  3. Approved AI Tools
  4. Prohibited Uses
  5. Data Handling Requirements
  6. Human Oversight Requirements
  7. Roles and Responsibilities
  8. Compliance and Consequences
  9. Exception Process
  10. Review and Updates

Section-by-Section Guide

1. Purpose and Scope

Sample Text:

This policy governs the use of Artificial Intelligence (AI) tools by all employees, contractors, and third parties acting on behalf of [Company Name]. It applies to all AI tools, including but not limited to generative AI (ChatGPT, Claude, Gemini), AI features in existing software, and specialised AI applications.

The purpose is to enable productive AI use while protecting the company, our customers, and our employees from AI-related risks.

2. Definitions

Include:

  • AI and Artificial Intelligence
  • Generative AI
  • Personal Information
  • Sensitive Information
  • Company Data
  • Approved Tools
  • Shadow AI

Example:

“Generative AI” means AI systems that create new content, including text, images, code, or other outputs, based on user prompts. This includes ChatGPT, Claude, Gemini, Copilot, and similar tools.

3. Approved AI Tools

List specifically approved tools and their permitted uses:

ToolApproved UsesRestrictions
Microsoft CopilotDocument drafting, email assistance, Excel analysisNo external customer data without de-identification
Xero AI (JAX)Financial queries, invoice managementNormal accounting data only, per Xero agreement
Canva Magic StudioMarketing content, internal documentsNo confidential information in prompts
[Your CRM] AI featuresLead scoring, email templatesWithin existing CRM data governance

Key Principle: If it’s not on the approved list, it’s not approved.

4. Prohibited Uses

Be specific about what’s forbidden:

Absolutely Prohibited:

  • Entering personal customer information into non-approved AI tools
  • Using AI for decisions that significantly affect individuals without human review
  • Submitting confidential business information to public AI tools
  • Using AI to generate misleading or deceptive content
  • Claiming AI-generated work as original human creation where disclosure is required
  • Using AI in ways that violate other company policies

Prohibited Unless Specifically Authorised:

  • AI for customer-facing communications without human review
  • AI for financial reporting or compliance documents
  • AI for legal documents or advice
  • AI for hiring or performance decisions

5. Data Handling Requirements

Categories:

Category A: Never input to any AI tool

  • Customer health or medical information
  • Complete customer databases
  • Financial account details
  • Employee personal files
  • Passwords or access credentials

Category B: Approved AI tools only

  • Customer contact information
  • Transaction history
  • Business financial data
  • Internal communications

Category C: Any approved AI tool

  • Publicly available information
  • De-identified data
  • Generic business content
  • Non-sensitive internal information

6. Human Oversight Requirements

The “Human-in-the-Loop” Principle:

All AI outputs intended for external communication or significant business decisions must be reviewed by a qualified human before use.

Specific Requirements:

  • Customer communications: Review by employee familiar with the customer relationship
  • Financial analysis: Review by someone qualified to assess accuracy
  • Legal content: Review by qualified professional before reliance
  • Technical code: Testing and review before deployment
  • Public content: Brand and accuracy review before publication

Documentation:

  • For significant decisions, document that human review occurred
  • Maintain audit trail of AI assistance and human oversight

7. Roles and Responsibilities

All Employees:

  • Comply with this policy
  • Report suspected violations
  • Complete AI awareness training
  • Ask if uncertain about AI use

Managers:

  • Ensure team compliance
  • Model appropriate AI use
  • Address violations promptly
  • Support training participation

AI Coordinator/Champion (designate someone):

  • Maintain approved tools list
  • Evaluate new AI tools
  • Answer policy questions
  • Monitor for Shadow AI
  • Recommend policy updates

Leadership:

  • Set AI strategy
  • Resource AI governance
  • Review policy annually
  • Address significant incidents

8. Compliance and Consequences

Monitoring: The company may monitor AI tool usage through [network monitoring/software audits/regular reviews] to ensure policy compliance.

Violations: Policy violations will be addressed through the standard disciplinary process, with consequences ranging from additional training to termination depending on severity and intent.

Reporting: Employees who identify potential policy violations should report to [AI Coordinator/Manager/HR] without fear of retaliation.

9. Exception Process

When Exceptions May Be Granted:

  • Pilot testing of new AI tools
  • Specific project requirements
  • Time-limited experiments

Process:

  1. Submit exception request to AI Coordinator
  2. Document proposed use, data involved, and risk mitigation
  3. Receive written approval before proceeding
  4. Report outcomes at conclusion

10. Review and Updates

This policy will be reviewed at least annually, or sooner if significant AI developments warrant. Employees will be notified of material changes.

Creating and implementing your AI Policy

Implementing the Policy

Phase 1: Preparation (Week 1-2)

  1. Audit current AI usage: Survey employees (anonymously if needed) about what AI tools they use
  2. Assess risks: Identify highest-risk current activities
  3. Select approved tools: Choose tools that meet security and compliance requirements
  4. Draft policy: Use framework above, customise for your business

Phase 2: Communication (Week 3-4)

  1. Leadership announcement: CEO/owner communicates AI governance importance
  2. Policy distribution: Share policy with all employees
  3. Q&A session: Address questions and concerns
  4. Training: Conduct initial AI awareness training

Phase 3: Enablement (Week 5-8)

  1. Deploy approved tools: Make sanctioned AI accessible
  2. Provide guidance: Create quick-reference materials for common tasks
  3. Establish support: Make AI Coordinator available for questions
  4. Monitor adoption: Track uptake of approved tools vs. shadow AI

Phase 4: Enforcement (Ongoing)

  1. Regular monitoring: Check for policy compliance
  2. Address violations: Handle consistently, focusing on education
  3. Gather feedback: Learn what’s working and what’s not
  4. Iterate: Update policy and approved tools based on experience

Transition from AI Chaos to AI Governance

Common Implementation Challenges

Challenge: “But I’m More Productive with ChatGPT”

Response: Validate the productivity benefit while explaining the risk. Offer an approved alternative. If no approved alternative exists, evaluate adding one.

Key Point: Your goal is safe AI use, not no AI use.

Challenge: “This Is Too Bureaucratic”

Response: Start with a minimal viable policy. Complex governance can grow over time. A simple policy followed is better than a comprehensive policy ignored.

Challenge: “We Don’t Have Resources for AI Governance”

Response: Governance doesn’t require a team. Designate existing staff with partial responsibility. Use templates (like this article) to accelerate policy creation.

Challenge: “Leadership Doesn’t Understand AI Risks”

Response: Present specific scenarios relevant to your business. Focus on regulatory, reputational, and financial risks leadership already cares about. Connect AI governance to existing risk management.

Challenge: “Employees Will Just Ignore This”

Response: Combine policy with enablement. Provide good alternatives to Shadow AI. Create positive incentives for appropriate use. Make compliance easy.

Measuring Success

Leading Indicators

  • Percentage of employees completing AI training
  • Adoption rate of approved AI tools
  • Policy acknowledgement rate
  • Questions/requests to AI Coordinator

Lagging Indicators

  • Shadow AI incidents detected
  • Policy violations requiring action
  • Security incidents involving AI
  • Productivity measures in AI-enabled functions

Annual Review Questions

  1. Has the approved tools list kept pace with needs?
  2. Are employees finding the policy practical?
  3. Have we had incidents that reveal policy gaps?
  4. Are we realising productivity benefits from AI?
  5. Is our maturity level improving?

Resources

Policy Template

Download our AI Acceptable Use Policy template at CloudGeeks Resources (link to your actual resource).

Training Materials

  • [Company] AI Quick Start Guide
  • Data Classification Reference
  • AI Tool Comparison Chart
  • Incident Reporting Form

External Resources

  • OAIC: Australian Privacy Principles guidelines
  • AIIA: Australian AI Ethics Framework
  • Deloitte: State of AI in the Enterprise reports

The Business Case for AI Governance

Investment in AI governance delivers:

Risk Reduction:

  • Lower likelihood of privacy breaches
  • Reduced regulatory exposure
  • Decreased legal liability

Productivity Gains:

  • Employees confident using approved tools
  • Less time wasted on risky workarounds
  • Clearer decision-making about AI use

Competitive Advantage:

  • Faster safe AI adoption than uncontrolled competitors
  • Ability to bid on governance-conscious contracts
  • Foundation for advanced AI applications

Cultural Benefits:

  • Innovation within boundaries
  • Trust between employees and leadership
  • Transparency about technology use

Conclusion

The “Human-in-the-Loop” isn’t about restricting AI—it’s about enabling it responsibly. Without governance, AI use happens anyway, just unsafely. With governance, you channel AI’s benefits while managing its risks.

Start simple:

  1. Assess your current AI maturity
  2. Create a basic acceptable use policy
  3. Designate someone to own AI governance
  4. Provide approved alternatives to shadow AI
  5. Iterate based on experience

The goal is progress, not perfection. A Level 2 organisation with basic governance is far safer than a Level 1 organisation where anything goes.

Ready to implement AI governance in your organisation? Contact CloudGeeks for assistance developing policies, selecting approved tools, and training your team on responsible AI use. We’ve helped numerous Australian businesses move from AI chaos to AI governance.

Keep humans in the loop. That’s where the wisdom lives.


Ready to transform your business?

Let's discuss how AI and cloud solutions can drive your digital transformation. Our team specializes in helping Australian SMBs implement cost-effective technology solutions.

Bella Vista, Sydney