Issue #29 · AI Agent Insider

EY Deploys Agents to 130,000 Auditors as Governance Gap Widens

Table of Contents

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

The Hook

Two things happened this week that should sit uncomfortably next to each other: EY handed 130,000 auditors a live multi-agent system on Tuesday, and researchers at RIT proved on the same day that agents built on Claude, Gemini, and DeepSeek will happily mishandle Social Security numbers if you let them. The gap between deployment velocity and governance readiness has never been wider – or more consequential.

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

This Week’s Signal

EY Goes All-In: 130,000 Auditors, One Multi-Agent Framework

On April 7, EY launched a global multi-agent AI framework baked directly into EY Canvas – the assurance platform that processes over 1.4 trillion lines of journal entry data per year across 160,000 audit engagements in more than 150 countries. The system launches with a core assistant and three specialized agents covering roughly 20 modular capabilities. EY’s stated target: 100% agent-supported audit activities by 2028.

The architecture runs on Microsoft Azure, Microsoft Foundry, and Microsoft Fabric. Notably, EY is treating the junior-staff disruption problem as a product problem – pairing platform agents with adaptive training to reskill entry-level auditors whose traditional learning path (doing the grunt work manually) just got automated.

Why this matters beyond Big Four: EY just ran the deployment playbook that every regulated-industry operator will study. The core pattern is: embed agents into an existing platform practitioners already trust, scope initial capabilities tightly (20 modular tasks, not open-ended autonomy), integrate governance from the infrastructure layer up, and build reskilling in parallel with rollout. The instinct to bolt governance on after deployment is exactly what fails. EY is engineering it in from day one.

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

3 Operator Playbooks

1. Agent 365 Goes GA – The Governance Stack Just Got a Name

Microsoft’s Agent 365 reached general availability on April 8, the same day MYOB announced a five-year partnership using it to serve 3.28 million Australian and New Zealand SMEs. Agent 365 is the orchestration and governance layer sitting above Copilot Studio and Microsoft Foundry – it handles policy enforcement, agent monitoring, and operational control across multi-agent environments. Reply was announced as a launch partner.

Your move: If you are deploying agents inside Microsoft’s ecosystem – or evaluating whether to – Agent 365 is now the governance foundation to benchmark your current approach against. The MYOB deal shows the pattern: Foundry for customer-facing agent deployment, Copilot Studio for internal agent creation, Agent 365 for the control plane. Understand the stack before you build around it.

2. Google Gemma 4 Makes On-Device Agentic AI Real

Google released Gemma 4 on April 2 under Apache 2.0: four models spanning an Effective 2B edge model up to a 31B dense model, with a 26B mixture-of-experts variant that activates only 3.8B parameters per token. All support native function calling, 140+ languages, audio and video inputs. Available immediately on Hugging Face, Kaggle, Ollama, and Google AI Studio.

Your move: The Apache 2.0 license is the operative word. You can run a capable tool-calling model behind your own firewall, on company hardware, with no per-seat fee and no need to renegotiate acceptable-use terms when a vendor updates its policy. Pick your lowest-risk internal workflow – document summarization, expense categorization, policy Q&A – and pilot the 4B or 26B MoE model locally before wiring it to any system that touches production data.

3. AudAgent: Your Agents May Be Leaking Data You Did Not Authorize

RIT cybersecurity researchers published AudAgent, a real-time compliance monitor that watches what AI agents actually do with sensitive data versus what their privacy policies say they will do. The results: agents powered by Claude, Gemini, and DeepSeek failed to refuse handling Social Security numbers through third-party tool integrations. GPT-4o consistently refused. The paper was accepted at the 2026 Privacy Enhancing Technologies Symposium.

Your move: Before wiring any agent to tools that touch user data – email, CRM, HR systems, payment platforms – run your own version of this test. Feed a synthetic SSN through the workflow and trace what happens. Most agents lack explicit refusal logic at the tool-call layer; they inherit whatever the underlying model was trained to do, which varies by vendor and version. Log every tool call, not just the final output. If you cannot audit it, you cannot defend it.

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Steal This

Agent Governance Pre-Flight Checklist

Before wiring any agent to a production system, verify each of the following:

[ ] Data scope defined: what data can the agent read, write, and transmit?
[ ] Tool-call logging enabled: every third-party call recorded with timestamp + payload hash
[ ] Sensitive data refusal tested: SSN, PII, credentials fed through synthetic test before launch
[ ] Human checkpoint defined: what actions require approval before execution?
[ ] Rollback path documented: how do you undo the last 10 agent actions?
[ ] Privacy policy alignment checked: does agent behavior match stated policy?
[ ] Incident escalation owner named: who gets paged when the agent acts unexpectedly?
[ ] Context window scope limited: agent sees only what it needs for its task

Adapt this for your stack. The CSA’s April 2026 survey of 1,500+ security leaders found 92% concerned about agent security – and most still lack a formal checklist like this one.

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

The Bottom Line

The week of April 8, 2026 is the week enterprise agentic AI stopped being theoretical. EY deployed to 130,000 professionals. Microsoft shipped a governance platform to market. Google put a function-calling agent on your laptop for free. And researchers handed you proof that the trust assumptions baked into most agent frameworks are wrong. The operators who will win the next 12 months are not the ones who moved fastest – they are the ones who moved fast and logged everything.


AI Agent Insider is published by Digital Forge Studios Inc.

Support the forge

Ko-fi Patreon
ETH0x3a4289F5e19C5b39353e71e20107166B3cCB2EDB BTC16Fhg23rQdpCr14wftDRWEv7Rzgg2qsj98 DOGEDNofxUZe8Q5FSvVbqh24DKJz6jdeQxTv8x