Back to jobs
AI Governance Lead (SDLC & Agentic AI Focus)
- Posted25 February 2026
- Salary$1000 - $1200 per day
- LocationSydney
- Job type Contract
- Expertise Sirius Technology
- ReferenceBH-66466
Job Description
AI Governance Lead (SDLC & Agentic AI Focus)
Reports To: Managing Director – AI Governance / Responsible AI
Role Overview As organisations accelerate adoption of Agentic AI and autonomous systems, this role ensures every AI-enabled solution is safe, compliant, and production-ready before deployment.
This is a hands-on AI Governance Lead position embedded directly into the Software Development Lifecycle (SDLC). You will operationalise Responsible AI principles across design, development, testing, deployment, and monitoring — implementing Shift-Left Governance to identify risk before code reaches production.
You will work closely with AI engineering squads, model risk teams, security, and senior stakeholders to integrate Governance-as-Code frameworks, automate compliance controls, and oversee adversarial testing of AI systems.
Core Responsibilities 1. Design & Ideation (Guardrails)
Reports To: Managing Director – AI Governance / Responsible AI
Role Overview As organisations accelerate adoption of Agentic AI and autonomous systems, this role ensures every AI-enabled solution is safe, compliant, and production-ready before deployment.
This is a hands-on AI Governance Lead position embedded directly into the Software Development Lifecycle (SDLC). You will operationalise Responsible AI principles across design, development, testing, deployment, and monitoring — implementing Shift-Left Governance to identify risk before code reaches production.
You will work closely with AI engineering squads, model risk teams, security, and senior stakeholders to integrate Governance-as-Code frameworks, automate compliance controls, and oversee adversarial testing of AI systems.
Core Responsibilities 1. Design & Ideation (Guardrails)
- Lead AI Impact Assessments during the requirements and architecture phase
- Define allowable AI use cases and ethical boundaries
- Ensure models comply with regulatory expectations and Responsible AI principles
- Assess risk in agentic / autonomous AI workflows
- Embed automated governance checks into IDEs (e.g., VS Code, Cursor) and CI/CD pipelines
- Implement Governance-as-Code frameworks to reduce manual review
- Scan for hallucinated libraries, insecure AI-generated code, and prohibited data usage
- Integrate automated policy enforcement into MLOps / LLMOps workflows
- Oversee adversarial testing and red-teaming workstreams
- Conduct jailbreak testing to assess data leakage and prompt exploitation risks
- Validate models for bias, hallucinations, and security vulnerabilities
- Ensure robust Human-in-the-Loop (HITL) controls
- Govern agent-to-agent communication and multi-agent behaviour
- Establish model drift, bias, and performance monitoring dashboards
- Define kill switches and rollback protocols for production AI
- Ensure audit-ready model lineage, traceability, and observability
- Provide oversight across the full AI lifecycle, from pilot to production
- Experience in AI governance, model risk, data governance, or technology risk
- Strong technical capability embedding governance into SDLC pipelines
- Hands-on experience with MLOps / LLMOps platforms such as Microsoft Azure Machine Learning or Amazon SageMaker
- Familiarity with adversarial AI testing tools such as Giskard, Deepchecks, or Microsoft Counterfit
- Practical knowledge of the NIST AI Risk Management Framework and ISO/IEC 42001
- Experience governing agentic AI / multi-agent systems
- Ability to translate regulatory and ethical principles into technical enforcement controls
- Time-to-Trust: Reduce time for AI pilots to pass ethical and risk review
- Governance Automation %: Increase percentage of automated compliance checks within CI/CD
- Risk Mitigation: Zero High or Critical findings in post-deployment AI audits
- Production Resilience: Effective drift detection and safe rollback in live environments