Field Guide
AI & Governance

Chapter 14: AI Governance & The Orchestrator Role

Building the Enterprise AI Governance Framework — privacy, security, and privilege — Human-in-the-Loop as operational discipline, and the rise of the Legal Engineer bridging law and data science.

Governance Is Critical

The deployment of AI in legal contexts creates a governance imperative that is qualitatively different from previous technology adoptions. A CLM that misroutes a contract creates an operational inconvenience. An AI system that generates correct legal advice, protects privileged information, and makes decisions aligned with regulatory requirements delivers competitive advantage and institutional trust.

AI governance in 2026 is a board-level concern. The EU AI Act imposes explicit obligations on organisations deploying AI in high-risk contexts (which includes significant legal decision-making). Australia's evolving AI governance framework creates expectations of transparency and accountability that early adopters can establish as competitive differentiators. The GC who implements robust AI governance builds institutional confidence, demonstrates control, and positions the organisation as a responsible AI leader.

The Enterprise AI Governance Framework

The Three Pillars

An effective AI governance framework for legal operations rests on three pillars: Privacy, Security, and Legal Professional Privilege (LPP). Each pillar requires specific controls, policies, and operational practices.

Pillar 1: Privacy

AI systems in legal process personal data — party names, contact details, employment information, health data in litigation contexts, financial data in transactional contexts. Privacy obligations under GDPR, the Australian Privacy Act, and equivalent regimes apply to AI processing with the same force as manual processing.

Data minimisation. AI systems should process only the personal data necessary for their function. A contract summarisation tool does not need access to the full text of every exhibit — it needs the commercial terms and key obligations. Design the data pipeline to exclude unnecessary personal data from AI processing.

Purpose limitation. Data processed by AI for one purpose (e.g., contract review) should not be repurposed for another (e.g., marketing analytics) without a separate legal basis. This is particularly relevant when AI vendors seek to use client data for model training — a practice that must be contractually prohibited or explicitly consented to.

Data subject rights. Individuals whose data is processed by legal AI systems retain their rights under applicable privacy legislation — including the right to access, correction, and deletion. The AI governance framework must include a process for identifying and responding to data subject requests that affect AI-processed data.

AI-specific privacy impact assessments. Before deploying any AI system that processes personal data, conduct a privacy impact assessment (PIA) that specifically addresses AI-related risks: automated profiling, inferential analytics, and the potential for AI outputs to reveal information about individuals that was not explicitly provided as input.

Pillar 2: Security

Model security. AI systems require protection against adversarial attacks — inputs specifically designed to manipulate the model's output. In a legal context, this means ensuring that a counterparty cannot craft contract language that causes the AI review tool to misclassify risky clauses. The governance framework should include testing for adversarial robustness and monitoring for anomalous AI behaviour.

Data security. Legal data processed by AI includes some of the most sensitive information in the enterprise — M&A plans, litigation strategy, regulatory investigation details, board deliberations. The AI system's data pipeline must meet or exceed the security classification of the most sensitive data it processes.

Access control. AI systems should enforce the same access controls as the underlying data they process. A junior lawyer who can access an AI system trained on partner-level files gains unintended access to restricted information. Role-based access control must extend to AI interfaces, ensuring data visibility restrictions flow through to AI access.

Vendor security assessment. Third-party AI tools must be assessed against the organisation's security standards before deployment. Key areas: data hosting location and sovereignty, encryption (in transit and at rest), sub-processor chains, incident response commitments, and data retention/deletion policies.

LPP presents the most legally complex governance challenge in AI deployment. Privilege protects communications between a lawyer and client made for the dominant purpose of obtaining or providing legal advice. Introducing AI into this communication chain creates novel questions:

Does privilege attach to AI-generated legal analysis? The answer varies by jurisdiction and is evolving in many. The robust position: treat AI-generated legal analysis as privileged and apply privilege protections consistently — marking outputs as privileged, restricting access, and maintaining the communication within the lawyer-client relationship.

Does inputting privileged information into an AI system waive privilege? If the AI system is operated by a third party with access to input data, privilege risk exists through disclosure to a non-privileged party. Protection: use AI systems that contractually commit to not accessing, reviewing, or retaining privileged data. Deploy on-premise or private-cloud instances where technically feasible to maintain exclusive control.

Does the AI system itself need to be covered by privilege? Many jurisdictions expect that the AI tool be treated as an agent of the lawyer (analogous to a paralegal or expert) for privilege to attach to its outputs. Structure the engagement accordingly: the AI tool should be deployed under the direction and control of a qualified lawyer, and its outputs should be reviewed and adopted (or modified) by that lawyer before being communicated to the client.

Warning

LPP risk is the most under-assessed risk in legal AI deployment. Organisations that process privileged materials through tools without privilege safeguards create vulnerability to opposing counsel challenges. This risk is actionable and material. Deploy AI tools that protect privilege only after implementing appropriate safeguards in your data architecture and vendor contracts.

Human-in-the-Loop (HITL): Operational Discipline

The Editor-in-Chief Model

HITL is not a checkbox — it is an operational discipline that defines how human oversight is structured at every stage of the AI workflow. The analogy is the newspaper editor-in-chief: journalists (the AI) produce content, but the editor reviews, fact-checks, and approves before publication. No content reaches the audience without editorial oversight.

HITL Design Principles

Define the review scope. Different AI outputs require appropriate review intensity based on risk. A low-risk NDA summary needs a 2-minute verification. A high-risk regulatory analysis requires thorough, clause-by-clause review. The HITL framework should specify review depth for each AI use case, calibrating effort to risk.

Assign qualified reviewers. The reviewer must be competent to assess the AI's output. An associate qualified to produce M&A due diligence reports can meaningfully review AI-generated analysis at that level. Match reviewer seniority and expertise to output complexity to ensure credible review.

Enable efficient review. Provide reviewers with tools that make verification fast and effective: cited sources for every AI assertion, confidence scores highlighting uncertain outputs, diff views showing what the AI changed or generated versus the source material. These tools accelerate human review and strengthen verification quality.

Track review outcomes. Log every HITL review: the AI output, the reviewer's assessment, any corrections made, and the final approved version. This data serves three purposes — it provides an audit trail, it generates training data for AI improvement, and it enables performance monitoring of both the AI and the reviewer.

Calibrating Autonomy Over Time

As AI systems demonstrate consistent accuracy, the intensity of human oversight can be gradually reduced — but never eliminated. The calibration follows a maturity curve:

StageHITL LevelCriteria
PilotReview 100% of outputsEstablishing baseline accuracy
Early ProductionReview 100%, fast-track confirmed-accurate categoriesAccuracy >90% sustained over 60 days
Mature ProductionReview flagged outputs + statistical sampleAccuracy >95% sustained over 6 months
OptimisedReview exceptions + quarterly auditAccuracy >98% sustained over 12 months

At every stage, "Optimised" includes active oversight. Even at the highest maturity level, the framework includes quarterly audits, exception review, and rapid escalation protocols when performance dips below thresholds.

A New Professional Archetype

The Legal Ops 3.0 era demands a professional who does not yet have an established career path: the Legal Engineer. This is the individual who bridges the gap between legal domain expertise and data/technology fluency — who can translate a lawyer's requirements into a system specification, configure an AI workflow, interpret analytics outputs, and communicate the results in both legal and technical language.

DomainRequired Competency
LegalUnderstanding of legal concepts, workflows, privilege, and risk — sufficient to specify requirements and evaluate AI outputs (not necessarily a practising lawyer)
DataData normalisation, taxonomy design, SQL or equivalent querying, basic statistical literacy — sufficient to prepare and maintain the data foundation AI depends on
TechnologyAPI integration concepts, workflow automation (no-code/low-code), CLM and matter management configuration, prompt engineering — sufficient to build and maintain AI-powered workflows
Change ManagementStakeholder engagement, training design, adoption measurement — sufficient to drive the behavioural change that technology deployment requires
Project ManagementAgile methodology, vendor management, budget tracking — sufficient to deliver initiatives on time and within scope

The Legal Engineer does not emerge fully formed. They evolve from one of three pathways:

Pathway 1: Lawyer → Legal Ops → Legal Engineer. A practising lawyer who transitions into Legal Ops, develops technology and data skills, and evolves into the bridging role. Strength: deep legal domain expertise. Investment area: deliberate technology skill development.

Pathway 2: Technologist → Legal Tech → Legal Engineer. A software engineer, data analyst, or IT professional who specialises in legal technology. Strength: strong technical foundation. Investment area: legal domain knowledge and substantive AI evaluation capability.

Pathway 3: Paralegal → Legal Ops → Legal Engineer. A paralegal with process expertise and operational knowledge who develops both legal and technical skills. Strength: deep understanding of how legal work actually flows. Investment area: seniority and strategic influence.

The most effective Legal Ops teams in 2026 include representation from at least two of these pathways, creating complementary skill coverage.

Strategic Insight

The Legal Engineer is the connective tissue that enables lawyers and technologists to work together productively. This role translates legal requirements into technical specifications and translates technical constraints into legal solutions. Strong Legal Engineer teams produce tools that work both technically and operationally, delivering intended business outcomes.

In the Trenches

The Privilege Leak

A multinational corporation deployed a generative AI tool for litigation support. The tool was excellent at summarising deposition transcripts, identifying inconsistencies across witness statements, and drafting preliminary legal memoranda. Lawyers on the litigation team adopted it enthusiastically.

Six months into deployment, opposing counsel in a major dispute issued a discovery request that specifically included "all documents and communications generated by or through artificial intelligence systems in connection with this litigation." The corporation's outside counsel reviewed the AI system's data architecture and discovered that the tool — a cloud-hosted SaaS platform — stored all inputs and outputs on shared infrastructure, with the vendor retaining access to the data for "product improvement" purposes. The vendor's terms of service, buried in the enterprise agreement, granted the vendor a licence to use input data for model training.

The opposing counsel's challenge: by providing privileged litigation strategy materials to a third-party vendor with ongoing access and usage rights, the corporation had potentially waived privilege over those materials. The court's ruling — delivered after four months of motion practice and $850K in legal fees — found privilege maintained for the specific documents at issue but signalled that future cases with more explicit vendor data usage might face different outcomes.

The corporation immediately migrated the tool to a private-cloud deployment with contractual terms prohibiting any vendor access to input or output data. They also implemented a mandatory privilege review checkpoint in the AI workflow: before any privileged content is processed through the AI system, a lawyer confirms that the data pipeline maintains privilege protections.

The episode required nearly $1M in direct legal fees and generated a privilege precedent affecting their future litigation strategy. The lesson: governance frameworks must assess both AI tool functionality and data handling to protect privilege throughout the AI lifecycle.

The Monday Morning Checklist

  • Audit your AI tools for privilege risk. For every AI tool currently in use by the legal team, answer: Does the vendor have access to input data? Does the vendor retain output data? Do the vendor's terms permit data usage for model training? Document privilege protection status for tools processing confidential content.
  • Draft (or update) your AI governance policy. The policy should cover: approved tools, prohibited tools, data classification rules, HITL requirements by use case, privacy impact assessment triggers, and incident response procedures. This policy is your highest-priority governance deliverable.
  • Define your HITL checkpoints. For each AI-powered workflow, specify: who reviews the output, what depth of review is required, what happens when the reviewer identifies an error, and how review outcomes are logged. Document this in an operational procedure that guides daily work.
  • Identify your Legal Engineer candidate. Look across your legal and legal ops team for the individual who most naturally bridges legal and technical domains. Invest in their development through formal training in data literacy, prompt engineering, or workflow automation, as this role will define the effectiveness of your AI strategy.