Exposure Brief
Issue 1

March 18, 2026

Executive Summary

Eighty-five percent of enterprises have deployed AI, but only 25% can see what their employees are doing with it, according to Optro’s March 2026 Risk Intelligence Report. That gap between adoption and visibility is widening. NIST’s AI 800-4 report describes post-deployment AI monitoring practices as “nascent.” Meanwhile, 73% of organizations have AI tools in production, and only 7% enforce policies on them in real time, per the Cybersecurity Insiders 2026 AI Security Report. The people being asked to close this gap are the same ones reading this briefing.

Lead Story

You Were Just Asked to Audit Every AI Tool in Your Organization. Now What?

“Leadership wants a full audit of every AI tool being used across the org. I genuinely don’t know how to produce one.”

That post on r/sysadmin (a community of 850,000+ IT professionals) pulled 523 upvotes and 218 comments when it landed on March 10, suggesting the problem is widespread among IT professionals. The mandate came down. The tools to execute it did not.

As the poster described it, the AI tools IT can inventory (managed licenses, enterprise subscriptions) are not the ones creating risk. As the poster described, the risk lives in personal ChatGPT accounts on managed devices, browser extensions routing inputs to AI backends, and employees using AI tools on personal phones to process work documents over mobile data. As the poster noted, corporate DLP (Data Loss Prevention), SWG (Secure Web Gateway), and CASB (Cloud Access Security Broker) each monitor a different layer. None of them monitors the prompt, and this structural gap is corroborated by Nightfall AI’s finding that traditional DLP achieves only 5-25% accuracy on AI browser data exfiltration pathways.

Optro’s 2026 Risk Intelligence Report, published March 17, puts the number at 85% of enterprises running AI in core operations, but only 25% with comprehensive visibility into how employees use it. Eighty percent described shadow AI as moderate to pervasive. The Cybersecurity Insiders 2026 AI Security Report, released March 16 and surveying 1,253 cybersecurity professionals, found an even starker gap: 73% have deployed AI tools, but only 7% enforce security policies in real time. Most cannot even distinguish a personal AI account from a corporate one.

No validated methodology for monitoring prompt-based data flow exists yet, according to NIST AI 800-4, published this month after three federal workshops and input from 200+ experts. The report says it plainly: best practices for monitoring AI after deployment are “nascent.” NIST’s findings align with what many IT teams have experienced firsthand, and NIST AI 800-4 gives you the citation to put in front of leadership.

Supporting Intelligence

Your DLP Policy Has a Prompt-Shaped Hole in It

A March 12 r/sysadmin thread details an attempt to write a DLP policy for AI interactions that ended in structural failure. Traditional DLP was built around file-based data movement: attachments, uploads, downloads. As Nightfall AI’s analysis found, copy-paste into a browser text field bypasses most existing DLP layers, with legacy solutions achieving only 5-25% accuracy on these pathways. As the poster put it: “SWG sees the domain. CASB sees the app. Neither sees the prompt.” Until DLP solutions include prompt-content inspection, the gap the poster described is likely to persist. Ask yours for a delivery timeline.

77% of Employees Are Pasting Company Data into AI Tools. On Personal Accounts.

Breached.Company’s Data Privacy Week 2026 analysis (February 2, 2026) found that 77% of employees paste company information into AI and LLM services, and 82% of them do it on personal accounts, not enterprise-managed tools. They’re pasting in email drafts, confidential negotiations, financial reports, customer data, and source code. For organizations relying on enterprise-managed tools as their data governance boundary, these numbers suggest the actual boundary may be elsewhere. The Cybersecurity Insiders 2026 AI Security Report corroborates the gap from the controls side: 92% of organizations lack semantic DLP (the kind that evaluates meaning, not pattern-matching), and 46% would fail to detect AI-rephrased sensitive content entirely.

The Federal Government Signed Off Without Seeing It

FedRAMP spent 480 hours over three years reviewing Microsoft’s Government Community Cloud High, ProPublica’s investigation revealed. Conducted 18 technical deep-dive sessions. Still couldn’t verify its fundamental security posture.

ProPublica’s investigation found that Microsoft failed to provide standard encryption documentation. Internal reviewers described the authorization package as “a pile of shit,” ProPublica reported. A former NSA computer scientist called it “security theater.” They approved it anyway, because agencies were already using it, ProPublica reported. FedRAMP reviewers checked compliance documents. They did not run independent scans of Microsoft’s encryption implementation. And the program doing the reviewing? FedRAMP now operates with roughly 24 staff and a $10M annual budget, after DOGE personnel cuts, per ProPublica. When an AI platform called Woflow got breached this month, the class action was filed in 10 days.

Your Audit Needs to Include the Agents

The visibility gap isn’t just about employees using AI tools. It’s about AI tools using your systems. The Cybersecurity Insiders 2026 AI Security Report found that AI agents (software that takes actions autonomously inside enterprise systems, sending emails, modifying records, querying databases without a human approving each step) are already operating inside more than half of organizations surveyed.

The report, released March 16, found that 56% of organizations have real agentic AI exposure. Twenty-three percent operate shadow agents they don’t know about. Most cannot stop an agent mid-action (91%), and over half have granted agents write access to collaboration tools. If the audit does not enumerate agents with write access, the next incident response will begin with a tool no one authorized.

Regulatory Radar

NIST AI 800-4 (Published March 2026): The first federal framework addressing post-deployment AI monitoring. Acknowledges that validated methodologies do not yet exist, which means the standards are being written now, and organizations already mapping to them may set the baseline. This is likely to become an audit reference point. Read the full report (PDF).

The Bottom Line