Agentic browsers — Atlas, Comet, Dia — are already on devices in your organisation. Some are managed. Many are not. Employees are installing them for the same reason they adopted every other productivity tool before it became official: it makes their work easier and no-one told them not to.
The question is whether you govern that adoption or let shadow AI answer it for you. Shadow AI is already embedded in how people work, and ungoverned agentic browsers mean zero telemetry, zero policy, and no incident response capability. That is a worse outcome than controlled deployment.
This playbook gives you a four-stage framework — Assess, Pilot, Govern, Monitor — built for companies with 50–500 employees and no dedicated SOC. It covers an acceptable use policy template, sandboxed pilot design, least-privilege configuration, SIEM telemetry integration, human-in-the-loop controls, shadow AI discovery, and a numbered incident response playbook for browser-agent compromise. It is part of a broader series on browser-agent security overview; the attack surface this playbook addresses — prompt injection and link-mediated exfiltration — is covered in the companion article.
Why is blocking agentic browsers not a viable strategy for most companies?
Blocking outright requires full endpoint control. And most companies with 50–500 employees simply do not have it. BYOD devices, contractor laptops, and unmanaged endpoints create gaps that no EDR or secure web gateway can close without enterprise-grade MDM — and most lean IT functions have not deployed it. Obrela recommends blocking agentic browser binaries for regulated sectors, but acknowledges that approach requires exactly the endpoint control most SMBs lack.
The deeper problem with blanket prohibition is what happens when it fails. Employees who adopt agentic browsers without governance produce the worst possible outcome: no telemetry, no logging, no policy guardrails, and no incident response capability. Cisco researchers found that 26% of agent skills analysed contained at least one security vulnerability — and those are the skills employees are pulling from unvetted registries.
The practical answer is controlled adoption with governance gates. Zero Trust architecture frames this well: treat every agent action as potentially untrusted, apply least privilege, and minimise the blast radius of any compromise. The incidents this controlled approach would have mitigated are real.
The first stage — assessing your current exposure — starts with knowing what is already installed. The discovery methods in the shadow AI section below are your starting point. For context on the full threat landscape this playbook is designed to address, the pillar article in this series maps all six attack surfaces.
How do you structure a sandboxed pilot before company-wide rollout?
A sandboxed pilot is a time-bounded deployment to a small, well-defined user cohort using non-production data and isolated system access. The goal is to validate your governance controls before you expose the broader organisation. Diana Kelley from Noma Security puts it well: “Don’t try to boil the ocean: start small, learn fast, and let the guardrails evolve.”
Here is how to structure it.
Who is in the pilot. Pick 5–10 people from a low-risk function — research, marketing, communications. Technically confident enough to report unexpected behaviour, but not connected to production systems or regulated data. Giskard’s analysis of Atlas explicitly recommends isolated testing environments with non-production data.
What the pilot environment must not touch. No access to internal applications holding customer PII, financial records, or production databases. Enterprise-managed options (Edge for Business, Island) are worth considering here — they provide centrally managed policy controls and MDM integration that consumer agentic browsers lack, which matters a lot when your IT function is lean.
Telemetry before the first user session. Get logging in place before anyone opens the browser. Agent activations, action types, cross-domain navigations, and content uploads must all be captured from day one. Retrofitting telemetry after deployment creates blind spots. Do not do it.
The phased timeline. Run weeks 1–4 in restricted read-only agent mode with no autonomous actions. From weeks 5–8, allow low-risk agent actions with human-in-the-loop confirmation on everything else. Disable agent/action mode by default — explicit opt-in per task, not a persistent session.
Exit criteria. Define these before the pilot starts. Close the pilot when: zero uncontrolled data exfiltration events have occurred, telemetry coverage is 100%, and all pilot users have completed AUP training — not on a calendar date. Tooling to support these controls is covered in the companion article.
How do you write an AI browser acceptable use policy for your engineering team?
An AUP for agentic browsers needs to exist before rollout, not after your first incident. A generic “AI usage policy” will not cover it. SentinelOne makes this point directly: confirm your policy addresses autonomous agents, not just chatbots.
Keep it to two pages maximum. Extend your existing IT acceptable use policy with agentic browser-specific clauses rather than building a parallel document. Nobody reads parallel documents.
Here is a template structure with eight numbered policy clauses:
-
Scope. Which agentic browser products are covered. Which user groups are authorised. Which devices are permitted. If a product is not on the approved list, it is not sanctioned.
-
Agent-mode activation. Agent/action mode must be explicitly activated per task. The default state is disabled. No persistent agent-mode sessions.
-
Prohibited uses. No automation of internal production systems without written approval. No agent access to customer PII, PHI, or PCI data outside the sandboxed pilot. No connecting agents to financial transaction systems.
-
Data handling. No pasting of classified or regulated data into agent prompts. No uploading internal documents to agent memory. Clear browser memory after sessions involving sensitive workflows.
-
Human oversight. All high-risk actions require human confirmation before execution. Users must not leave agent sessions unattended during autonomous workflows.
-
Logging and audit. All agent sessions must generate telemetry. Users must not disable or circumvent logging. Sessions are subject to periodic audit review.
-
Incident reporting. Any unexpected agent behaviour must be reported immediately through your existing incident reporting channel. Users must not attempt to resolve agent misbehaviour themselves.
-
Policy violations. Align consequences with your existing IT AUP. First violation triggers a review. Repeated violations result in access revocation.
Review this policy quarterly — a policy written in Q1 can have gaps by Q3 as new products and capabilities arrive. Set update triggers for material new releases or major public incidents. Compliance frameworks these controls satisfy are covered in the compliance article in this series.
What are the specific steps to configure least-privilege access for agentic browser features?
Least-privilege access means disabling everything by default and enabling capabilities per workflow. Simple as that. The more permissions the agent has, the larger the blast radius of any compromise. Giskard’s analysis of Atlas confirms that in agent mode, a successfully injected instruction can propagate across multiple authenticated sessions and domains before detection.
Five specific configuration steps:
-
Disable autonomous action by default. Require explicit opt-in per task. No persistent agent mode.
-
Restrict agent scope to approved domains. No open-web navigation during agentic sessions. If the agent does not need to visit a domain to complete the task, it should not be able to.
-
Block agent access to local file system, environment variables, and developer tools. Most productivity use cases do not require these, and they are common targets.
-
Enforce enterprise SSO and MFA for any agent session that touches internal systems. Consumer agentic browsers often retain data indefinitely and rely on users to manage privacy settings — do not depend on user-managed defaults.
-
Separate agentic browsing from sensitive browsing. Do not allow agent-mode sessions to run in the same context as sessions accessing production systems or regulated data.
There is also a DLP gap worth flagging. Most data loss prevention tools were built for files and emails, not dynamic browser action content. PII, PHI, and PCI data flowing through agent prompts is a leakage vector that traditional DLP misses entirely. Extend DLP controls to cover agent prompt inputs, not just file uploads.
How do you integrate agentic browser telemetry into your existing SIEM or XDR setup?
Agentic browser telemetry is the stream of structured events — agent activations, action types, data uploads, cross-domain navigations — that feeds your monitoring and detection stack. You do not need an enterprise SIEM to get started. Lightweight open-source options like Wazuh or Security Onion work fine. The requirement is simply that telemetry is collected at all.
Six specific signals to capture:
-
Agent activation events. When agent/action mode is turned on, by which user, on which device, at what time.
-
Action type logging. What the agent did — form submission, navigation, file upload, API call — with timestamps and target URLs.
-
Cross-domain navigation. When the agent navigates from an approved domain to an unapproved one. High-priority alert signal.
-
Content upload volume. Large or unusual data uploads during agent sessions.
-
Human-in-the-loop bypass events. Any override or dismissal of a confirmation prompt during a high-risk agent action.
-
Session duration anomalies. Unusually long sessions, or sessions outside business hours.
Four example alert rules:
- Agent session initiated on an unmanaged device accessing an internal domain — severity: high.
- Cross-domain navigation from internal portal to external URL during an agent session — severity: medium.
- Content upload exceeding 5MB during an agent session — severity: medium.
- Agent session active for more than 60 minutes without human interaction — severity: low, escalate if combined with other signals.
Noma Security’s Agentic Risk Map visualises all AI agents in an organisation — useful alongside your SIEM. Zenity’s Correlation Agent surfaces manipulation attempts and reduces investigation time.
When should you require human-in-the-loop confirmation and how do you configure it?
Human-in-the-loop (HiTL) controls require explicit human confirmation before the agent executes high-risk or irreversible actions. Treating this as a binary toggle — on for everything, or off for everything — either paralyses productivity or removes oversight entirely. You need a tiered decision framework.
hCaptcha tested five major browser agents on 20 common abuse scenarios and found these products attempted nearly every malicious request with no jailbreaking required. Vendor-native HiTL controls are immature. So use external controls — browser extension policies, proxy rules, SIEM alerts — to enforce confirmation requirements until vendors build more granular options.
A three-tier classification:
Always require confirmation (high-risk): Financial transactions. Form submissions to external services. Access to internal systems with customer data. Credential entry or OAuth authorisation. Any first-time action in the pilot.
Require confirmation on first use, then allow (medium-risk): Navigation to approved external domains. Content summarisation of non-sensitive pages. File downloads from known sources.
Allow unattended (low-risk): Read-only browsing of public web content. Search queries. Navigation within a single approved domain.
“High-risk” means anything that could result in data exposure, financial loss, or a compliance issue if the agent acts incorrectly. Without a dedicated SOC, that definition should be inclusive — when in doubt, require confirmation.
Schedule quarterly reviews of what sits in the “allow unattended” tier. Agents gain capabilities, and autonomy creep is a real problem — actions that were safe six months ago quietly stop being safe. A quarterly review keeps your HiTL controls calibrated. Compliance frameworks these controls satisfy are addressed in the compliance article in this series.
How do you discover and govern shadow AI browser adoption that is already underway?
Shadow AI discovery is not a one-time exercise. Even after rollout, new tools appear on devices without going through the adoption process. You cannot govern what you cannot see. Zenity automatically identifies Atlas, Comet, Dia, and other agentic tools across managed and unmanaged devices if you want automated discovery. If you do not have an enterprise endpoint agent, here are five lightweight methods.
-
Network proxy and firewall log inspection. Look for outbound connections to known agentic browser API endpoints — api.perplexity.ai, chatgpt.com agent endpoints, Dia browser service domains. Even a basic firewall can log these.
-
Browser extension audit. Query managed devices for installed browser extensions associated with agentic browsing capabilities. Chrome enterprise management or basic UEM can enumerate extensions without additional tooling.
-
UEM/MDM device inventory. Query for installed applications matching known agentic browser binaries — Comet, Atlas, Dia, Fellou — on managed devices.
-
User survey. Direct outreach asking what AI tools teams are using. Position this as governance enablement, not an investigation — honest responses come when people believe the goal is to help them use tools safely.
-
DNS query monitoring. Monitor DNS requests for agentic browser service domains. Lightweight, does not require deep packet inspection, and surfaces both managed and unmanaged device activity.
When you find shadow AI, the governance response matters as much as the discovery. The goal is to transition from unsanctioned to sanctioned usage, not to punish early adopters. Employees who self-adopted are often your most productivity-motivated people — alienating them does not serve anyone.
Create a self-service adoption request process: if an employee wants to use an agentic browser, submission triggers a vendor due-diligence check and AUP acknowledgement. This channels shadow AI into the governance framework. How prompt injection creates these risks is covered in the prompt injection article in this series.
What to do in the first hour if a browser agent is prompt-injected in production
This is a numbered incident response playbook, not narrative framing. Prompt injection is unlikely to ever be fully solved — OpenAI acknowledges this about Atlas. What you can control is response speed and damage containment. Real-world examples of these risks illustrate what this playbook addresses.
Step 1: Isolate the agent session immediately
Terminate the active agent session — close the browser, revoke the session token, or kill the process.
For managed browsers (Edge for Business, Island), use the admin console to remotely terminate. For unmanaged browsers (Atlas, Comet), instruct the user to close all windows and disconnect from the network.
Do not continue using the browser to investigate. The agent may still be acting on the injected prompt.
Step 2: Review forensic logs from the agent session
Pull telemetry from your SIEM or XDR for the compromised session: action log, navigation history, content uploads, form submissions.
Identify what the agent did — which domains it navigated to, what data it accessed, what it uploaded or submitted. Look for indicators of prompt injection: unexpected external navigation, form submissions the user did not initiate, data uploads to unfamiliar endpoints.
Step 3: Clear AI memory and persistent context
If the agentic browser retains persistent memory across sessions — Atlas and Dia both do — clear the agent’s memory store. An injected prompt can persist into future sessions if memory is not purged.
Delete browser memory, conversation history, and cached context associated with the compromised session. Verify the clearing was successful — some products require multiple steps or admin-level access to fully purge persistent context.
Step 4: Rotate credentials for all sessions the agent touched
Priority order: (1) OAuth tokens or API keys the agent used; (2) SSO session tokens; (3) passwords for any service the agent logged into; (4) shared credentials the agent may have accessed.
Scope this to every system the agent navigated to, plus any system where those credentials provide access. If the agent operated within enterprise SSO, invalidate the user’s session and force re-authentication across all connected services.
Step 5: Identify all data the agent accessed during the compromised session
Map the exposure: open tabs, page content, local files (if file system access was available), clipboard contents, form field data, and internal application screens. Zenity Labs notes that agentic browsers often hold deep access to local files, cloud sessions, and developer environments — the scope can be wide. This inventory feeds directly into Step 6.
Step 6: Determine whether compliance notification is required
GDPR trigger: if the agent accessed or exfiltrated personal data of EU residents, a 72-hour breach notification clock may start — consult legal counsel.
HIPAA trigger: if the agent accessed protected health information, breach notification obligations apply.
SOC 2 trigger: if the agent accessed systems in scope for a SOC 2 audit, document the incident for the next audit cycle.
Without a compliance officer in-house, engage external legal counsel if any regulated data was in scope. GDPR and SOC 2 obligations are covered in the compliance article in this series. Obrela’s compliance framing for browser-agent incidents is a useful reference.
Step 7: Conduct a post-incident review and red-team the injection vector
Identify the specific prompt injection vector — malicious webpage, email, document, or another external input.
Reproduce the injection in the sandboxed environment to confirm mitigation. Update the AUP and HiTL controls based on what the incident revealed. Brief the team without blame — prompt hygiene training requires real examples to stick. Open-source tools for red teaming and detection can support the exercise.
Frequently Asked Questions
What is an agentic browser and how is it different from a regular browser with AI features?
An agentic browser embeds an LLM directly into the browsing environment, enabling it to autonomously navigate, submit forms, and execute workflows. A regular browser with AI add-ons provides suggestions and summaries but does not take autonomous action. That distinction matters. The agent can take actions the user did not explicitly initiate, across any authenticated session it can reach — which is a fundamentally different class of risk.
Can my company be held liable if a browser agent leaks customer data because of a prompt injection?
Potentially yes. GDPR and HIPAA apply regardless of whether a human or an AI agent took the action — the organisation is still the data controller. Giskard’s analysis of Atlas recommends treating consumer agentic browsers as out of scope for regulated data until enterprise-grade audit and identity management are in place. Detailed treatment of liability is in the compliance article in this series.
How long should a sandboxed pilot run before we consider company-wide rollout?
At least eight weeks — four of restricted read-only usage followed by four of graduated autonomy with HiTL controls. The pilot closes when exit criteria are met (zero exfiltration events, full telemetry coverage, all users AUP-trained), not on a calendar date.
Do I need an enterprise SIEM to monitor agentic browser activity?
No. Lightweight open-source options like Wazuh or Security Onion work fine. Even basic syslog forwarding is better than no monitoring. The requirement is that telemetry is collected at all. Enterprise platforms add correlation and automation, but they are not prerequisites.
What is the difference between an agentic browser and an enterprise secure browser like Island or Edge for Business?
Enterprise secure browsers (Island, Microsoft Edge for Business) provide central policy control, MDM integration, and administrative audit capabilities. Consumer agentic browsers prioritise AI capability and user experience but lack enterprise governance features — policy management, audit logging, and MDM all fall back on your IT team, which adds real overhead to a controlled pilot.
How do I prevent employees from using agentic browsers on personal devices for work tasks?
You cannot fully prevent it on unmanaged devices. Make the sanctioned path easier instead: provide a governed option through the pilot programme, a self-service adoption request process, and an AUP with clear expectations. DNS and proxy monitoring gives partial visibility.
What does “prompt hygiene” mean and how do I train my team on it?
Prompt hygiene is awareness of what context, data, and instructions go into the AI agent — the same concept as phishing awareness training applied to LLM interactions. Noma Security’s CISO guide covers the core points: never paste sensitive data into agent prompts, verify agent actions before confirming, and report unexpected behaviour immediately.
How do I know if an agentic browser has been compromised?
Check your telemetry for: unexpected cross-domain navigation, form submissions the user did not initiate, data uploads to unfamiliar endpoints, sessions outside business hours, and OAuth token generation linked to non-standard applications. The SIEM integration section above provides specific alert rules.
Should I evaluate consumer agentic browsers or enterprise secure browsers for our pilot?
Start with an enterprise-managed option (Edge for Business, Island) if central policy control and MDM integration are priorities. Evaluate consumer agentic browsers (Atlas, Comet) only within the sandboxed pilot and only after governance controls are in place.
What is the minimum viable security stack for governing agentic browsers in an SMB?
At minimum: (1) an AUP covering agent-mode usage, (2) basic telemetry via firewall logs or lightweight SIEM, (3) DNS monitoring for known agentic browser endpoints, (4) browser extension audit capability, and (5) an incident response playbook. No enterprise EDR, CASB, or SASE required. Start with what you have and build from there.