<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0">
    <channel>
        <title>VentureBeat</title>
        <link>https://venturebeat.com/feed/</link>
        <description>Transformative tech coverage that matters</description>
        <lastBuildDate>Sun, 12 Apr 2026 22:00:21 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>https://github.com/jpmonette/feed</generator>
        <language>en</language>
        <copyright>Copyright 2026, VentureBeat</copyright>
        <item>
            <title><![CDATA[Five signs data drift is already undermining your security models]]></title>
            <link>https://venturebeat.com/security/five-signs-data-drift-is-already-undermining-your-security-models</link>
            <guid isPermaLink="false">7i0FSJc3DXou1wEJSVEmkG</guid>
            <pubDate>Sun, 12 Apr 2026 19:00:00 GMT</pubDate>
            <description><![CDATA[<p>Data drift happens when the statistical properties of a machine learning (ML) model&#x27;s input data change over time, eventually rendering its predictions less accurate. <a href="https://venturebeat.com/security/ocsf-explained-the-shared-data-language-security-teams-have-been-missing?_gl=1*yt0z35*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">Cybersecurity professionals</a> who rely on ML for tasks like malware detection and network threat analysis find that undetected data drift can create vulnerabilities. A model trained on old attack patterns may fail to see today&#x27;s sophisticated threats. Recognizing the early signs of data drift is the first step in maintaining reliable and efficient security systems.</p><h2><b>Why data drift compromises security models</b></h2><p>ML models are trained on a snapshot of historical data. When live data no longer resembles this snapshot, the model&#x27;s performance dwindles, creating a <a href="https://venturebeat.com/technology/why-cios-must-lead-ai-experimentation-not-just-govern-it?_gl=1*x7qiq4*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">critical cybersecurity risk</a>. A threat detection model may generate more false negatives by missing real breaches or create more false positives, leading to alert fatigue for security teams.</p><p>Adversaries actively exploit this weakness. In 2024,<a href="https://thehackernews.com/2024/07/proofpoint-email-routing-flaw-exploited.html"> <u>attackers used echo-spoofing techniques</u></a> to bypass email protection services. By exploiting misconfigurations in the system, they sent millions of spoofed emails that evaded the vendor&#x27;s ML classifiers. This incident demonstrates how threat actors can manipulate input data to exploit blind spots. When a security model fails to adapt to shifting tactics, it becomes a liability.</p><h2><b>5 indicators of data drift</b></h2><p>Security professionals can recognize the presence of drift (or its potential) in several ways.</p><h3><b>1. A sudden drop in model performance</b></h3><p>Accuracy, precision, and recall are often the first casualties. A consistent decline in these key metrics is a red flag that the model is no longer in sync with the current threat landscape.</p><p>Consider Klarna&#x27;s success: Its AI assistant handled 2.3 million customer service conversations in its first month and performed work equivalent to 700 agents. This efficiency drove a<a href="https://www.nutshell.com/blog/best-ai-chatbots"> <u>25% decline in repeat inquiries</u></a> and reduced resolution times to under two minutes. </p><p>Now imagine if those parameters suddenly reversed because of drift. In a security context, a similar drop in performance does not just mean unhappy clients — it also means successful intrusions and potential data exfiltration.</p><h3><b>2. Shifts in statistical distributions</b></h3><p><a href="https://venturebeat.com/security/human-centric-iam-is-failing-agentic-ai-requires-a-new-identity-control?_gl=1*61shbb*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">Security teams</a> should monitor the core statistical properties of input features, such as the mean, median, and standard deviation. A significant change in these metrics from training data could indicate the underlying data has changed.</p><p>Monitoring for such shifts enables teams to catch drift before it causes a breach. For example, a phishing detection model might be trained on emails with an average attachment size of 2MB. If the average attachment size suddenly jumps to 10MB due to a new malware-delivery method, the model may fail to classify these emails correctly.</p><h3><b>3. Changes in prediction behavior</b></h3><p>Even if overall accuracy seems stable, distributions of predictions might change, a phenomenon often referred to as prediction drift.</p><p>For instance, if a fraud detection model historically flagged 1% of transactions as suspicious but suddenly starts flagging 5% or 0.1%, either something has shifted or the nature of the input data has changed. It might indicate a new type of attack that confuses the model or a change in legitimate user behavior that the model was not trained to identify.</p><h3><b>4. An increase in model uncertainty</b></h3><p>For models that provide a confidence score or probability with their predictions, a general decrease in confidence can be a subtle sign of drift.</p><p>Recent studies highlight the<a href="https://arxiv.org/html/2410.21952v2"> <u>value of uncertainty quantification</u></a> in detecting adversarial attacks. If the model becomes less sure about its forecasts across the board, it is likely facing data it was not trained on. In a cybersecurity setting, this uncertainty is an early sign of potential model failure, suggesting the model is operating in unfamiliar ground and that its decisions might no longer be reliable.</p><h3><b>5. Changes in feature relationships</b></h3><p>The correlation between different input features can also change over time. In a network intrusion model, traffic volume and packet size might be highly linked during normal operations. If that correlation disappears, it can signal a change in network behavior that the model may not understand. A sudden feature decoupling could indicate a new tunneling tactic or a stealthy exfiltration attempt.</p><h2><b>Approaches to detecting and mitigating data drift</b></h2><p>Common detection methods include the Kolmogorov-Smirnov (KS) and the population stability index (PSI). These compare the <a href="https://towardsdatascience.com/drift-detection-in-robust-machine-learning-systems/"><u>distributions of live and training data</u></a> to identify deviations. The KS test determines if two datasets differ significantly, while the PSI measures how much a variable&#x27;s distribution has shifted over time. </p><p>The mitigation method of choice often depends on how the drift manifests, as distribution changes may occur suddenly. For example, customers&#x27; buying behavior may change overnight with the launch of a new product or a promotion. In other cases, drift may occur gradually over a more extended period. That said, security teams must learn to adjust their monitoring cadence to capture both rapid spikes and slow burns. Mitigation will involve retraining the model on more recent data to reclaim its effectiveness.</p><h2><b>Proactively manage drift for stronger security</b></h2><p>Data drift is an inevitable reality, and cybersecurity teams can maintain a strong security posture by treating detection as a continuous and automated process. Proactive monitoring and model retraining are fundamental practices to ensure ML systems remain reliable allies against developing threats.</p><p><i>Zac Amos is the Features Editor at </i><a href="https://rehack.com/"><i><u>ReHack</u></i></a><i>.</i></p>]]></description>
            <category>Security</category>
            <category>DataDecisionMakers</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/erAw6FrOeAX9eZJqeF2Dx/3a759d02f32a698bdc815c787701a17a/AI_drift.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Your developers are already running AI locally: Why on-device inference is the CISO’s new blind spot]]></title>
            <link>https://venturebeat.com/security/your-developers-are-already-running-ai-locally-why-on-device-inference-is</link>
            <guid isPermaLink="false">3EC5GemarqXB92UGk1xUjb</guid>
            <pubDate>Sun, 12 Apr 2026 15:00:20 GMT</pubDate>
            <description><![CDATA[<p>For the last 18 months, the CISO playbook for generative AI has been relatively simple: Control the browser.</p><p><a href="https://venturebeat.com/security/ocsf-explained-the-shared-data-language-security-teams-have-been-missing?_gl=1*4903t3*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">Security teams</a> tightened cloud access security broker (CASB) policies, blocked or monitored traffic to well-known AI endpoints, and routed usage through sanctioned gateways. The operating model was clear: If sensitive data leaves the network for an external API call, we can observe it, log it, and stop it. But that model is starting to break.</p><p>A quiet hardware shift is pushing large language model (LLM) usage off the network and onto the endpoint. Call it Shadow AI 2.0, or the “bring your own model” (BYOM) era: Employees running capable models locally on laptops, offline, with no API calls and no obvious network signature. The governance conversation is still framed as “data exfiltration to the cloud,” but the more immediate enterprise risk is increasingly “unvetted inference inside the device.&quot;</p><p>When inference happens locally, traditional data loss prevention (DLP) doesn’t see the interaction. And when security can’t see it, it can’t manage it.</p><h3><b>Why local inference is suddenly practical</b></h3><p>Two years ago, running a useful LLM on a work laptop was a niche stunt. Today, it’s routine for technical teams.</p><p>Three things converged:</p><ul><li><p><b>Consumer-grade accelerators got serious: </b>A MacBook Pro with 64GB unified memory can often run quantized 70B-class models at usable speeds (with practical limits on context length). What once required multi-GPU servers is now feasible on a high-end laptop for many real workflows.</p></li><li><p><b>Quantization went mainstream:</b> It’s now easy to compress models into smaller, faster formats that fit within laptop memory often with acceptable quality tradeoffs for many tasks.</p></li><li><p><b>Distribution is frictionless:</b> Open-weight models are a single command away, and the tooling ecosystem makes “download → run → chat” trivial.</p></li></ul><p><b>The result: </b>An engineer can pull down a multi‑GB model artifact, turn off Wi‑Fi, and run sensitive workflows locally, source code review, document summarization, drafting customer communications, even exploratory analysis over regulated datasets. No outbound packets, no proxy logs, no cloud audit trail.</p><p>From a <a href="https://venturebeat.com/security/mythos-detection-ceiling-security-teams-new-playbook?_gl=1*qe97gz*_up*MQ..*_ga*MzY1OTQzODYzLjE3NzYwMDU1Mjk.*_ga_SCH1J7LNKY*czE3NzYwMDU1MjgkbzEkZzAkdDE3NzYwMDU1MjgkajYwJGwwJGgw*_ga_B8TDS1LEXQ*czE3NzYwMDU1MjgkbzEkZzAkdDE3NzYwMDU1MjgkajYwJGwwJGgw">network-security perspective</a>, that activity can look indistinguishable from “nothing happened”.</p><h3><b>The risk isn’t only data leaving the company anymore</b></h3><p>If the data isn’t leaving the laptop, why should a CISO care?</p><p>Because the dominant risks shift from exfiltration to integrity, provenance, and compliance. In practice, local inference creates three classes of blind spots that most enterprises have not operationalized.</p><h4><b>1. Code and decision contamination (integrity risk)</b></h4><p>Local models are often adopted because they’re fast, private, and “no approval required.&quot; The downside is that they’re frequently unvetted for the enterprise environment.</p><p><b>A common scenario:</b> A senior developer downloads a community-tuned coding model because it benchmarks well. They paste in internal auth logic, payment flows, or infrastructure scripts to “clean it up.&quot; The model returns output that looks competent, compiles, and passes unit tests, but subtly degrades security posture (weak input validation, unsafe defaults, brittle concurrency changes, dependency choices that aren’t allowed internally). The engineer commits the change.</p><p>If that interaction happened offline, you may have no record that AI influenced the code path at all. And when you later do incident response, you’ll be investigating the symptom (a vulnerability) without visibility into a key cause (uncontrolled model usage).</p><h4><b>2. Licensing and IP exposure (compliance risk)</b></h4><p>Many high-performing models ship with licenses that include <a href="https://llama.meta.com/llama3/license/"><u>restrictions on commercial use</u></a>, attribution requirements, field-of-use limits, or obligations that can be incompatible with proprietary product development. When employees run models locally, that usage can bypass the organization’s normal procurement and legal review process.</p><p>If a team uses a non-commercial model to generate production code, documentation, or product behavior, the company can inherit risk that shows up later during M&amp;A diligence, customer security reviews, or litigation. The hard part is not just the license terms, it’s the lack of inventory and traceability. Without a governed model hub or usage record, you may not be able to prove what was used where.</p><h4><b>3. Model supply chain exposure (provenance risk)</b></h4><p>Local inference also changes the software supply chain problem. Endpoints begin accumulating large model artifacts and the toolchains around them: ownloaders, converters, runtimes, plugins, UI shells, and Python packages.</p><p>There is a critical technical nuance here: The file format matters. While newer formats like <a href="https://huggingface.co/docs/safetensors/index"><b><u>Safetensors</u></b></a> are designed to prevent arbitrary code execution, older <a href="https://pytorch.org/docs/stable/generated/torch.load.html"><b><u>Pickle-based</u></b><u> PyTorch files</u></a> can execute malicious payloads simply when loaded. If your developers are grabbing unvetted checkpoints from Hugging Face or other repositories, they aren&#x27;t just downloading data — they could be downloading an exploit.</p><p>Security teams have spent decades learning to treat unknown executables as hostile. BYOM requires extending that mindset to model artifacts and the surrounding runtime stack. The biggest organizational gap today is that most companies have no equivalent of a <a href="https://www.cisa.gov/sbom"><u>software bill of materials</u></a> for models: Provenance, hashes, allowed sources, scanning, and lifecycle management.</p><h3><b>Mitigating BYOM: treat model weights like software artifacts</b></h3><p>You can’t solve local inference by blocking URLs. You need endpoint-aware controls and a developer experience that makes the safe path the easy path.</p><p>Here are three practical ways:</p><p><b>1. Move governance down to the endpoint</b> </p><p>Network DLP and CASB still matter for cloud usage, but they’re not sufficient for BYOM. Start treating local model usage as an endpoint governance problem by looking for specific signals:</p><ul><li><p><b>Inventory and detection:</b> Scan for high-fidelity indicators like .gguf files larger than 2GB, processes like <a href="https://github.com/ggerganov/llama.cpp"><u>llama.cpp</u></a> or Ollama, and local listeners on common <a href="https://docs.ollama.com/faq"><u>default port 11434</u></a>.</p></li><li><p><b>Process and runtime awareness:</b> Monitor for repeated high GPU/NPU (neural processing unit) utilization from unapproved runtimes or unknown local inference servers.</p></li><li><p><b>Device policy:</b> Use <b>mobile device management (MDM) and endpoint detection and response (EDR)</b> policies to control installation of unapproved runtimes and enforce baseline hardening on engineering devices. The point isn’t to punish experimentation. It’s to regain visibility.</p></li></ul><p><b>2. Provide a paved road: An internal, curated model hub</b> </p><p><a href="https://venturebeat.com/security/ai-agent-zero-trust-architecture-audit-credential-isolation-anthropic-nvidia-nemoclaw">Shadow AI</a> is often an outcome of friction. Approved tools are too restrictive, too generic, or too slow to approve. A better approach is to offer a curated internal catalog that includes: </p><ul><li><p>Approved models for common tasks (coding, summarization, classification)</p></li><li><p>Verified licenses and usage guidance</p></li><li><p>Pinned versions with hashes (prioritizing safer formats like Safetensors)</p></li><li><p>Clear documentation for safe local usage, including where sensitive data is and isn’t allowed. If you want developers to stop scavenging, give them something better.</p></li></ul><p><b>3. Update policy language: “Cloud services” isn’t enough anymore</b> </p><p>Most acceptable use policies talk about SaaS and cloud tools. BYOM requires policy that explicitly covers:</p><ul><li><p>Downloading and running model artifacts on corporate endpoints</p></li><li><p>Acceptable sources</p></li><li><p>License compliance requirements</p></li><li><p>Rules for using models with sensitive data</p></li><li><p>Retention and logging expectations for local inference tools This doesn’t need to be heavy-handed. It needs to be unambiguous.</p></li></ul><h3><b>The perimeter is shifting back to the device</b></h3><p>For a decade we moved security controls “up” into the cloud. Local inference is pulling a meaningful slice of AI activity back “down” to the endpoint.</p><p>5 signals shadow AI has moved to endpoints:</p><ul><li><p><b>Large model artifacts:</b> Unexplained storage consumption by .gguf or .pt files.</p></li><li><p><b>Local inference servers:</b> Processes listening on ports like 11434 (Ollama).</p></li><li><p><b>GPU utilization patterns:</b> Spikes in GPU usage while offline or disconnected from VPN.</p></li><li><p><b>Lack of model inventory:</b> Inability to map code outputs to specific model versions.</p></li><li><p><b>License ambiguity:</b> Presence of &quot;non-commercial&quot; model weights in production builds.</p></li></ul><p>Shadow AI 2.0 isn’t a hypothetical future, it’s a predictable consequence of fast hardware, easy distribution, and developer demand. CISOs who focus only on network controls will miss what’s happening on the silicon sitting right on employees’ desks.</p><p>The next phase of AI governance is less about blocking websites and more about controlling artifacts, provenance, and policy at the endpoint, without killing productivity.</p><p><i>Jayachander Reddy Kandakatla is a senior MLOps engineer.</i></p>]]></description>
            <category>Security</category>
            <category>DataDecisionMakers</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/pAoHef9hMVI3aHoyHfluC/f410fef5dc2a910939184a98db76eec4/AI_perimeter.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[AI agent credentials live in the same box as untrusted code. Two new architectures show where the blast radius actually stops.]]></title>
            <link>https://venturebeat.com/security/ai-agent-zero-trust-architecture-audit-credential-isolation-anthropic-nvidia-nemoclaw</link>
            <guid isPermaLink="false">20HlmDVeOH0foEOboDiQiE</guid>
            <pubDate>Fri, 10 Apr 2026 22:26:53 GMT</pubDate>
            <description><![CDATA[<p>Four separate RSAC 2026 keynotes arrived at the same conclusion without coordinating. Microsoft&#x27;s Vasu Jakkal told attendees that zero trust must extend to AI. Cisco&#x27;s Jeetu Patel called for a shift from access control to action control, <a href="https://venturebeat.com/security/rsac-2026-agent-identity-frameworks-three-gaps">saying in an exclusive interview with VentureBeat</a> that agents behave &quot;more like teenagers, supremely intelligent, but with no fear of consequence.&quot; CrowdStrike&#x27;s George Kurtz identified AI governance as the biggest gap in enterprise technology. Splunk&#x27;s John Morgan called for an agentic trust and governance model. Four companies. Four stages. <a href="https://cloudsecurityalliance.org/blog/2026/04/03/every-rsac-keynote-asked-the-same-five-questions-here-s-the-framework-that-answers-them">One problem</a>.</p><p>Matt Caulfield, VP of Product for Identity and Duo at Cisco, put it bluntly in an exclusive VentureBeat interview at RSAC. &quot;While the concept of zero trust is good, we need to take it a step further,&quot; Caulfield said. &quot;It&#x27;s not just about authenticating once and then letting the agent run wild. It&#x27;s about continuously verifying and scrutinizing every single action the agent&#x27;s trying to take, because at any moment, that agent can go rogue.&quot;</p><p>Seventy-nine percent of organizations already use AI agents, according to <a href="https://www.pwc.com/us/en/tech-effect/ai-analytics/ai-agent-survey.html">PwC&#x27;s 2025 AI Agent Survey</a>. Only 14.4% reported full security approval for their entire agent fleet, per the <a href="https://www.gravitee.io/state-of-ai-agent-security">Gravitee State of AI Agent Security 2026 report</a> of 919 organizations in February 2026. A <a href="https://cloudsecurityalliance.org/artifacts/the-state-of-ai-security-and-governance">CSA survey</a> presented at RSAC found that only 26% have AI governance policies. <a href="https://cloudsecurityalliance.org/blog/2026/02/02/the-agentic-trust-framework-zero-trust-governance-for-ai-agents">CSA&#x27;s Agentic Trust Framework</a> describes the resulting gap between deployment velocity and security readiness as a governance emergency.</p><p>Cybersecurity leaders and industry executives at RSAC agreed on the problem. Then two companies shipped architectures that answer the question differently. The gap between their designs reveals where the real risk sits.</p><h2>The monolithic agent problem that security teams are inheriting</h2><p>The default enterprise agent pattern is a monolithic container. The model reasons, calls tools, executes generated code, and holds credentials in one process. Every component trusts every other component. OAuth tokens, API keys, and git credentials sit in the same environment where the agent runs code it wrote seconds ago.</p><p>A prompt injection gives the attacker everything. Tokens are exfiltrable. Sessions are spawnable. The blast radius is not the agent. It is the entire container and every connected service.</p><p>The <a href="https://cloudsecurityalliance.org/press-releases/2026/03/24/more-than-two-thirds-of-organizations-cannot-clearly-distinguish-ai-agent-from-human-actions">CSA and Aembit survey</a> of 228 IT and security professionals quantifies how common this remains: 43% use shared service accounts for agents, 52% rely on workload identities rather than agent-specific credentials, and 68% cannot distinguish agent activity from human activity in their logs. No single function claimed ownership of AI agent access. Security said it was a developer&#x27;s responsibility. Developers said it was a security responsibility. Nobody owned it.</p><p>CrowdStrike CTO Elia Zaitsev, in an exclusive VentureBeat interview, said the pattern should look familiar. &quot;A lot of what securing agents look like would be very similar to what it looks like to secure highly privileged users. They have identities, they have access to underlying systems, they reason, they take action,&quot; Zaitsev said. &quot;There&#x27;s rarely going to be one single solution that is the silver bullet. It&#x27;s a defense in depth strategy.&quot;</p><p>CrowdStrike CEO George Kurtz highlighted ClawHavoc (a supply chain campaign targeting the OpenClaw agentic framework) at RSAC during his <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">keynote</a>. <a href="https://www.koi.ai/blog/clawhavoc-341-malicious-clawedbot-skills-found-by-the-bot-they-were-targeting">Koi Security</a> named the campaign on February 1, 2026. Antiy CERT confirmed 1,184 malicious skills tied to 12 publisher accounts, according to <a href="https://vpncentral.com/clawhavoc-poisons-openclaw-clawhub-with-1184-malicious-skills/">multiple</a> <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">independent</a> analyses of the campaign. <a href="https://snyk.io/blog/toxicskills-malicious-ai-agent-skills-clawhub/">Snyk&#x27;s ToxicSkills research</a> found that 36.8% of the 3,984 ClawHub skills scanned contain security flaws at any severity level, with 13.4% rated critical. Average breakout time has dropped to 29 minutes. Fastest observed: 27 seconds. (<a href="https://www.crowdstrike.com/en-us/press-releases/2026-crowdstrike-global-threat-report/">CrowdStrike 2026 Global Threat Report</a>)</p><h2>Anthropic separates the brain from the hands</h2><p><a href="https://www.anthropic.com/engineering/managed-agents">Anthropic&#x27;s Managed Agents</a>, launched April 8 in public beta, split every agent into three components that do not trust each other: a brain (Claude and the harness routing its decisions), hands (disposable Linux containers where code executes), and a session (an append-only event log outside both).</p><p>Separating instructions from execution is one of the oldest patterns in software. Microservices, serverless functions, and message queues. </p><p>Credentials never enter the sandbox. Anthropic stores OAuth tokens in an external vault. When the agent needs to call an MCP tool, it sends a session-bound token to a dedicated proxy. The proxy fetches real credentials from the vault, makes the external call, and returns the result. The agent never sees the actual token. Git tokens get wired into the local remote at sandbox initialization. Push and pull work without the agent touching the credential. For security directors, this means a compromised sandbox yields nothing an attacker can reuse.</p><p>The security gain arrived as a side effect of a performance fix. Anthropic decoupled the brain from the hands so inference could start before the container booted. Median time to first token <a href="https://www.anthropic.com/engineering/managed-agents">dropped roughly 60%</a>. The zero-trust design is also the fastest design. That kills the enterprise objection that security adds latency.</p><p>Session durability is the third structural gain. A container crash in the monolithic pattern means total state loss. In Managed Agents, the session log persists outside both brain and hands. If the harness crashes, a new one boots, reads the event log, and resumes. No state lost turns into a productivity gain over time. Managed Agents include built-in session tracing through the <a href="https://www.anthropic.com/engineering/managed-agents">Claude Console</a>.</p><p>Pricing: $0.08 per session-hour of active runtime, idle time excluded, plus standard API token costs. Security directors can now model agent compromise cost per session-hour against the cost of the architectural controls.</p><h2>Nvidia locks the sandbox down and monitors everything inside it</h2><p><a href="https://github.com/NVIDIA/NemoClaw">Nvidia&#x27;s NemoClaw</a>, released March 16 in early preview, takes the opposite approach. It does not separate the agent from its execution environment. It wraps the entire agent inside four stacked security layers and watches every move. Anthropic and Nvidia are the only two vendors to have shipped zero-trust agent architectures publicly as of this writing; others are in development.</p><p>NemoClaw stacks five enforcement layers between the agent and the host. Sandboxed execution uses Landlock, seccomp, and network namespace isolation at the kernel level. Default-deny outbound networking forces every external connection through explicit operator approval via <a href="https://docs.nvidia.com/nemoclaw/latest/reference/network-policies.html">YAML-based policy</a>. Access runs with minimal privileges. A privacy router directs sensitive queries to locally-running Nemotron models, cutting token cost and data leakage to zero. The layer that matters most to security teams is intent verification: OpenShell&#x27;s policy engine intercepts every agent action before it touches the host. The trade-off for organizations evaluating NemoClaw is straightforward. Stronger runtime visibility costs more operator staffing.</p><p>The agent does not know it is inside NemoClaw. In-policy actions return normally. Out-of-policy actions get a configurable denial.</p><p>Observability is the strongest layer. A real-time Terminal User Interface logs every action, every network request, every blocked connection. The audit trail is complete. The problem is cost: operator load scales linearly with agent activity. Every new endpoint requires manual approval. Observation quality is high. Autonomy is low. That ratio gets expensive fast in production environments running dozens of agents.</p><p>Durability is the gap nobody&#x27;s talking about. Agent state persists as files inside the sandbox. If the sandbox fails, the state goes with it. No external session recovery mechanism exists. Long-running agent tasks carry a durability risk that security teams need to price into deployment planning before they hit production.</p><h2>The credential proximity gap</h2><p>Both architectures are a real step up from the monolithic default. Where they diverge is the question that matters most to security teams: how close do credentials sit to the execution environment?</p><p>Anthropic removes credentials from the blast radius entirely. If an attacker compromises the sandbox through prompt injection, they get a disposable container with no tokens and no persistent state. Exfiltrating credentials requires a two-hop attack: influence the brain&#x27;s reasoning, then convince it to act through a container that holds nothing worth stealing. Single-hop exfiltration is structurally eliminated.</p><p>NemoClaw constrains the blast radius and monitors every action inside it. Four security layers limit lateral movement. Default-deny networking blocks unauthorized connections. But the agent and generated code share the same sandbox. Nvidia&#x27;s privacy router keeps inference credentials on the host, outside the sandbox. But messaging and integration tokens (Telegram, Slack, Discord) are injected into the sandbox as runtime environment variables. Inference API keys are proxied through the privacy router and not passed into the sandbox directly. The exposure varies by credential type. Credentials are policy-gated, not structurally removed.</p><p>That distinction matters most for indirect prompt injection, where an adversary embeds instructions in content the agent queries as part of legitimate work. A poisoned web page. A manipulated API response. The intent verification layer evaluates what the agent proposes to do, not the content of data returned by external tools. Injected instructions enter the reasoning chain as trusted context. With proximity to execution.</p><p>In the Anthropic architecture, indirect injection can influence reasoning but cannot reach the credential vault. In the NemoClaw architecture, injected context sits next to both reasoning and execution inside the shared sandbox. That is the widest gap between the two designs.</p><p>NCC Group&#x27;s David Brauchler, Technical Director and Head of AI/ML Security, <a href="https://www.esecurityplanet.com/artificial-intelligence/rsac-2026-rethinking-trust-in-agentic-ai-security/">advocates for gated agent architectures</a> built on <a href="https://www.nccgroup.com/research/analyzing-secure-ai-architectures/">trust segmentation principles</a> where AI systems inherit the trust level of the data they process. Untrusted input, restricted capabilities. Both Anthropic and Nvidia move in this direction. Neither fully arrives.</p><h2>The zero-trust architecture audit for AI agents</h2><p>The audit grid covers three vendor patterns across six security dimensions, five actions per row. It distills to five priorities:</p><ol><li><p><b>Audit every deployed agent for the monolithic pattern. </b>Flag any agent holding OAuth tokens in its execution environment. The <a href="https://cloudsecurityalliance.org/press-releases/2026/03/24/more-than-two-thirds-of-organizations-cannot-clearly-distinguish-ai-agent-from-human-actions">CSA data</a> shows 43% use shared service accounts. Those are the first targets.</p></li><li><p><b>Require credential isolation in agent deployment RFPs. </b>Specify whether the vendor removes credentials structurally or gates them through policy. Both reduce risk. They reduce it by different amounts with different failure modes.</p></li><li><p><b>Test session recovery before production. </b>Kill a sandbox mid-task. Verify state survives. If it does not, long-horizon work carries a data-loss risk that compounds with task duration.</p></li><li><p><b>Staff for the observability model. </b>Anthropic&#x27;s console tracing integrates with existing observability workflows. NemoClaw&#x27;s TUI requires an operator-in-the-loop. The staffing math is different.</p></li><li><p><b>Track indirect prompt injection roadmaps. </b>Neither architecture fully resolves this vector. Anthropic limits the blast radius of a successful injection. NemoClaw catches malicious proposed actions but not malicious returned data. Require vendor roadmap commitments on this specific gap.</p></li></ol><p>Zero trust for AI agents stopped being a research topic the moment two architectures shipped. The monolithic default is a liability. The 65-point gap between deployment velocity and security approval is where the next class of breaches will start.</p>]]></description>
            <author>louiswcolumbus@gmail.com (Louis Columbus)</author>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/35u1BcqPcGPsUcDOjxBtIh/fb7d22c10100687068f227c644d2e297/zero_trust_hero.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Intuit compressed months of tax code implementation into hours — and built a workflow any regulated-industry team can adapt]]></title>
            <link>https://venturebeat.com/data/intuit-compressed-months-of-tax-code-implementation-into-hours-and-built-a</link>
            <guid isPermaLink="false">5uV2DLvz7YtlRiLwgWoGEL</guid>
            <pubDate>Fri, 10 Apr 2026 22:09:14 GMT</pubDate>
            <description><![CDATA[<p>When the <a href="https://www.congress.gov/bill/119th-congress/house-bill/1/text">One Big Beautiful Bill</a> arrived as a 900-page unstructured document — with no standardized schema, no published IRS forms, and a hard shipping deadline — Intuit&#x27;s TurboTax team had a question: could AI compress a months-long implementation into days without sacrificing accuracy?</p><p>What they built to do it is less a tax story than a template, a workflow combining commercial AI tools, a proprietary domain-specific language and a custom unit test framework that any domain-constrained development team can learn from.</p><p>Joy Shaw, director of tax at Intuit, has spent more than 30 years at the company and lived through both the <a href="https://www.irs.gov/newsroom/tax-cuts-and-jobs-act-a-comparison-for-businesses">Tax Cuts and Jobs Act</a> and the OBBB. &quot;There was a lot of noise in the law itself and we were able to pull out the tax implications, narrow it down to the individual tax provisions, narrow it down to our customers,&quot; Shaw told VentureBeat. &quot;That kind of distillation was really fast using the tools, and then enabled us to start coding even before we got forms and instructions in.&quot;</p><h2>How the OBBB raised the bar</h2><p>When the Tax Cuts and Jobs Act passed in 2017, the TurboTax team worked through the legislation without AI assistance. It took months, and the accuracy requirements left no room for shortcuts. </p><p>&quot;We used to have to go through the law and we&#x27;d code sections that reference other law code sections and try and figure it out on our own,&quot; Shaw said.</p><p>The OBBB arrived with the same accuracy requirements but a different profile. At 900-plus pages, it was structurally more complex than the TCJA. It came as an unstructured document with no standardized schema. The House and Senate versions used different language to describe the same provisions. And the team had to begin implementation before the IRS had published official forms or instructions.</p><p>The question was whether AI tools could compress the timeline without compromising the output. The answer required a specific sequence and tooling that did not exist yet.</p><h2>From unstructured document to domain-specific code</h2><p>The OBBB was still moving through Congress when the TurboTax team began working on it. Using large language models, the team summarized the House version, then the Senate version and then reconciled the differences. Both chambers referenced the same underlying tax code sections, a consistent anchor point that let the models draw comparisons across structurally inconsistent documents.</p><p>By signing day, the team had already filtered provisions to those affecting TurboTax customers, narrowed to specific tax situations and customer profiles. Parsing, reconciliation and provision filtering moved from weeks to hours.</p><p>Those tasks were handled by ChatGPT and general-purpose LLMs. But those tools hit a hard limit when the work shifted from analysis to implementation. TurboTax does not run on a standard programming language. Its tax calculation engine is built on a proprietary domain-specific language maintained internally at Intuit. Any model generating code for that codebase has to translate legal text into syntax it was never trained on, and identify how new provisions interact with decades of existing code without breaking what already works.</p><p>Claude became the primary tool for that translation and dependency-mapping work. Shaw said it could identify what changed and what did not, letting developers focus only on the new provisions.

&quot;It&#x27;s able to integrate with the things that don&#x27;t change and identify the dependencies on what did change,&quot; she said. &quot;That sped up the process of development and enabled us to focus only on those things that did change.&quot;</p><h2>Building tooling matched to a near-zero error threshold</h2><p>General-purpose LLMs got the team to working code. Getting that code to shippable required two proprietary tools built during the OBBB cycle.</p><p>The first auto-generated TurboTax product screens directly from the law changes. Previously, developers curated those screens individually for each provision. The new tool handled the majority automatically, with manual customization only where needed.</p><p>The second was a purpose-built unit test framework. Intuit had always run automated tests, but the previous system produced only pass/fail results. When a test failed, developers had to manually open the underlying tax return data file to trace the cause.

&quot;The automation would tell you pass, fail, you would have to dig into the actual tax data file to see what might have been wrong,&quot; Shaw said. The new framework identifies the specific code segment responsible, generates an explanation and allows the correction to be made inside the framework itself.</p><p>Shaw said accuracy for a consumer tax product has to be close to 100 percent. Sarah Aerni, Intuit&#x27;s VP of technology for the Consumer Group, said the architecture has to produce deterministic results. 

&quot;Having the types of capabilities around determinism and verifiably correct through tests — that&#x27;s what leads to that sort of confidence,&quot; Aerni said.</p><p>The tooling handles the speed. But Intuit also uses LLM-based evaluation tools to validate AI-generated output, and even those require a human tax expert to assess whether the result is correct. &quot;It comes down to having human expertise to be able to validate and verify just about anything,&quot; Aerni said.</p><h2>Four components any regulated-industry team can use</h2><p>The OBBB was a tax problem, but the underlying conditions are not unique to tax. Healthcare, financial services, legal tech and government contracting teams regularly face the same combination: complex regulatory documents, hard deadlines, proprietary codebases, and near-zero error tolerance.</p><p>Based on Intuit&#x27;s implementation, four elements of the workflow are transferable to other domain-constrained development environments:</p><ol><li><p><b>Use commercial LLMs for document analysis.</b> General-purpose models handle parsing, reconciliation and provision filtering well. That is where they add speed without creating accuracy risk.</p></li><li><p><b>Shift to domain-aware tooling when analysis becomes implementation.</b> General-purpose models generating code into a proprietary environment without understanding it will produce output that cannot be trusted at scale.</p></li><li><p><b>Build evaluation infrastructure before the deadline, not during the sprint.</b> Generic automated testing produces pass/fail outputs. Domain-specific test tooling that identifies failures and enables in-context fixes is what makes AI-generated code shippable.</p></li><li><p><b>Deploy AI tools across the whole organization, not just engineering.</b> Shaw said Intuit trained and monitored usage across all functions. AI fluency was distributed across the organization rather than concentrated in early adopters.</p></li></ol><p>&quot;We continue to lean into the AI and human intelligence opportunity here, so that our customers get what they need out of the experiences that we build,&quot; Aerni said.
</p>]]></description>
            <category>Data</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/6B3W5NVKiHfs0041gAHmty/091b174e5bac4070f7551204014de6d0/turbotax-ai-smk.jpg?w=300&amp;q=30" length="0" type="image/jpg"/>
        </item>
        <item>
            <title><![CDATA[OpenAI introduces ChatGPT Pro $100 tier with 5X usage limits for Codex compared to Plus]]></title>
            <link>https://venturebeat.com/orchestration/openai-introduces-chatgpt-pro-usd100-tier-with-5x-usage-limits-for-codex</link>
            <guid isPermaLink="false">4Ze71PAUykgaXOprdqyQ6x</guid>
            <pubDate>Thu, 09 Apr 2026 23:07:00 GMT</pubDate>
            <description><![CDATA[<p>OpenAI is making moves to try and court more developers and vibe coders (those who build software using AI models and natural language) away from rivals like Anthropic.</p><p>Today, the firm arguably most synonymous with the generative AI boom <a href="https://x.com/OpenAI/status/2042295688323875316">announced</a> it will begin offering a new, more mid-range subscription tier — a $100 ChatGPT Pro plan — which <a href="https://chatgpt.com/pricing/">joins</a> its free, Go ($8 monthly), Plus ($20 monthly) and existing Pro ($200 monthly) plans for individuals using ChatGPT and related OpenAI products.</p><p>OpenAI also currently offers Edu, Business ($25 per user monthly, formerly known as Team) and Enterprise (variably priced) plans for organizations in said sectors. </p><h2><b>Why offer a $100 monthly ChatGPT Pro plan?</b></h2><p>So why introduce a new $100 ChatGPT Pro plan, then? </p><p>The big selling point from OpenAI is that the new plan offers five times greater usage limits on Codex, the company&#x27;s agentic vibe coding application/harness (the name is shared by both, as well as a lineup of coding-specific language models), than the existing, $20 monthly Plus plan, which seems fair given the math ($20x5=$100). </p><p>As OpenAI co-founder and CEO Sam Altman <a href="https://x.com/sama/status/2042342572958630332">wrote in a post on X</a>: &quot;It is very nice to see Codex getting so much love. We are launching a $100 ChatGPT Pro tier by very popular demand.&quot;</p><p>However, alongside this, <a href="https://x.com/OpenAI/status/2042295690382995789">OpenAI&#x27;s official company account on X</a> noted that &quot;we’re rebalancing Codex usage in [ChatGPT] Plus to support more sessions throughout the week, rather than longer sessions in a single day.&quot;</p><p>That sounds a lot like OpenAI is also simultaneously <i>reducing</i> how much ChatGPT Plus users can use its Codex harness and application per day. </p><h2><b>What are the new usage limits for the new $100 ChatGPT Pro plan vs. the $20 Plus?</b></h2><p>So, what are the current limits on the $20 Plus plan? The new Pro plan gives you 5X greater than...what? </p><p>Turns out, this is trickier than you&#x27;d think to calculate, because it actually varies depending on which underlying AI model you are using to power the Codex application or harness, and whether you are working on code stored in the cloud or locally on your machine or servers. </p><p><a href="https://developers.openai.com/codex/pricing?codex-credit-costs=plus-pro-enterprise-edu-legacy&amp;codex-usage-limits=pro-20x#what-are-the-usage-limits-for-my-plan">OpenAI’s Developer website</a> underwent several updates today, so we&#x27;ve only reflected the latest pricing structure and offerings below as of Thursday, April at 10:45 pm ET. It notes that for individual users, Codex usage is categorized by “Local Messages” (tasks run on the user’s machine) and “Cloud Tasks” (tasks run on OpenAI’s infrastructure), and those limits share a five-hour rolling window. </p><p>It also says additional weekly limits may apply. The current Codex pricing page now shows lower displayed usage ranges than the older version, and it measures Code Reviews in a five-hour window rather than per week. For Pro 5x specifically, OpenAI says the currently shown limits include a temporary 2x usage boost that ends May 31, 2026.</p><h3><b>ChatGPT Plus ($20/month)</b></h3><ul><li><p><b>GPT-5.4: </b>20–100 local messages every 5 hours.</p></li><li><p><b>GPT-5.4-mini: </b>60–350 local messages every 5 hours.</p></li><li><p><b>GPT-5.3-Codex: </b>30–150 local messages and 10–60 cloud tasks every 5 hours.</p></li><li><p><b>Code Reviews:</b> 20–50 every 5 hours.</p></li></ul><h3><b>ChatGPT Pro 5x ($100/month)</b></h3><ul><li><p><b>GPT-5.4: </b>200–1,000 local messages every 5 hours.</p></li><li><p><b>GPT-5.4-mini:</b> 600–3,500 local messages every 5 hours.</p></li><li><p><b>GPT-5.3-Codex: </b>300–1,500 local messages and 100–600 cloud tasks every 5 hours.</p></li><li><p><b>Code Reviews: </b>200–500 every 5 hours.</p></li></ul><p><b>Note: </b>The limits shown for Pro 5x include a temporary 2x usage boost that ends May 31, 2026.</p><h3><b>ChatGPT Pro 20x ($200/month)</b></h3><ul><li><p><b>GPT-5.4: </b>400–2,000 local messages every 5 hours.</p></li><li><p><b>GPT-5.4-mini: </b>1,200–7,000 local messages every 5 hours.</p></li><li><p><b>GPT-5.3-Codex:</b> 600–3,000 local messages and 200–1,200 cloud tasks every 5 hours.</p></li><li><p><b>Code Reviews:</b> 400–1,000 every 5 hours.</p></li><li><p><b>Exclusive access:</b> Includes GPT-5.3-Codex-Spark in research preview for ChatGPT Pro users only. OpenAI says it has its own separate usage limit, which may adjust based on demand.</p></li></ul><p>And as<a href="https://help.openai.com/en/articles/11369540-using-codex-with-your-chatgpt-plan"> OpenAI&#x27;s Help documentation </a>states:</p><p><i>&quot;The number of Codex messages you can send within these limits varies based on the size and complexity of your coding tasks, and where you execute tasks. Small scripts or simple functions may only consume a fraction of your allowance, while larger codebases, long running tasks, or extended sessions that require Codex to hold more context will use significantly more per message.&quot; </i></p><h2><b>The larger strategic implications and context</b></h2><p>OpenAI’s sudden move toward the $100 price point and expanded agentic capacity comes amid the unprecedented financial ascent of its chief rival, Anthropic. </p><p>Just days ago, <a href="https://www.anthropic.com/news/google-broadcom-partnership-compute">Anthropic revealed its annualized run-rate revenue (ARR) has topped $30 billion</a>, surpassing <a href="https://finance.yahoo.com/news/openai-tops-25-billion-annualized-033836274.html">OpenAI&#x27;s last reported ARR of approximately $24–$25 billion. </a></p><p>This growth has been fueled by the massive adoption of Claude Code and Claude Cowork, products that have set the benchmark for enterprise-grade autonomous coding.</p><p>The competitive friction intensified on April 4, 2026, when <a href="https://venturebeat.com/technology/anthropic-cuts-off-the-ability-to-use-claude-subscriptions-with-openclaw-and">Anthropic officially blocked Claude subscriptions</a> from being used to provide the intelligence for third-party agentic AI harnesses like OpenClaw. </p><p>To be clear, Anthropic Claude models themselves can still be used with OpenClaw, users just must now pay for access to Claude models through Anthropic&#x27;s application programming interface (API) or extra usage credits, rather than as part of the monthly Claude subscription tiers (which some have likened to an &quot;all-you-can eat&quot; buffet, making the economics challenging for Anthropic when power users and third-party harnesses like OpenClaw consume more than the $20 or $200 monthly user spend on the plans in tokens). </p><p>OpenClaw’s creator, Peter Steinberger, was <a href="https://venturebeat.com/technology/openais-acquisition-of-openclaw-signals-the-beginning-of-the-end-of-the">notably hired by OpenAI in February 2026</a> to lead their personal agent strategy, and has, since joining, <a href="https://x.com/steipete/status/2040209434019082522">actively spoken out against Anthropic&#x27;s limitations</a> — advising that <a href="https://x.com/steipete/status/2039019069257756735">OpenAI&#x27;s Codex</a> and models generally don&#x27;t have the same restrictions as Anthropic is now imposing.</p><p>By hiring Steinberger and subsequently launching a Pro tier that provides the high-volume capacity Anthropic recently restricted, OpenAI is effectively courting the displaced OpenClaw community to reclaim the professional developer market.</p>]]></description>
            <author>carl.franzen@venturebeat.com (Carl Franzen)</author>
            <category>Orchestration</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/6uVTE16gVLv4XRlDrOmhO8/ddd73f327394bbdce9e8ac05e7114f06/ChatGPT_Image_Apr_9__2026__06_14_45_PM.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Mythos autonomously exploited vulnerabilities that survived 27 years of human review. Security teams need a new detection playbook]]></title>
            <link>https://venturebeat.com/security/mythos-detection-ceiling-security-teams-new-playbook</link>
            <guid isPermaLink="false">5HQfJFuyh5FqRQeSMq4Xla</guid>
            <pubDate>Thu, 09 Apr 2026 18:30:00 GMT</pubDate>
            <description><![CDATA[<p>A 27-year-old bug sat inside <a href="https://ftp.openbsd.org/pub/OpenBSD/patches/7.8/common/025_sack.patch.sig">OpenBSD’s TCP stack</a> while auditors reviewed the code, fuzzers ran against it, and the operating system earned its reputation as one of the most security-hardened platforms on earth. Two packets could crash any server running it. Finding that bug cost a single Anthropic discovery campaign approximately $20,000. The specific model run that surfaced the flaw cost under $50.</p><p>Anthropic’s Claude Mythos Preview found it. Autonomously. No human guided the discovery after the initial prompt.</p><h2>The capability jump is not incremental </h2><p>On Firefox 147 exploit writing, Mythos succeeded 181 times versus 2 for <a href="https://www.anthropic.com/news/claude-opus-4-6">Claude Opus 4.6</a>. A 90x improvement in a single generation. <a href="https://www.swebench.com/">SWE-bench</a> Pro: 77.8% versus 53.4%. <a href="https://red.anthropic.com/2026/mythos-preview/">CyberGym</a> vulnerability reproduction: 83.1% versus 66.6%. Mythos saturated Anthropic’s Cybench CTF at 100%, forcing the red team to shift to real-world zero-day discovery as the only meaningful evaluation left. Then it surfaced thousands of zero-day vulnerabilities across every major operating system and every major browser, many one to two decades old. Anthropic engineers with no formal security training asked Mythos to find remote code execution vulnerabilities overnight and woke up to a complete, working exploit by morning, according to <a href="https://red.anthropic.com/2026/mythos-preview/">Anthropic’s red team assessment</a>.</p><p>Anthropic assembled <a href="https://www.anthropic.com/glasswing">Project Glasswing</a>, a 12-partner defensive coalition including <a href="https://www.crowdstrike.com/">CrowdStrike</a>, <a href="https://www.cisco.com/">Cisco</a>, <a href="https://www.paloaltonetworks.com/">Palo Alto Networks</a>, Microsoft, AWS, Apple, and the <a href="https://www.linuxfoundation.org/">Linux Foundation</a>, backed by $100 million in usage credits and $4 million in open-source grants. Over 40 additional organizations that build or maintain critical software infrastructure also received access. The partners have been running Mythos against their own infrastructure for weeks. Anthropic committed to a public findings report “within 90 days,” landing in early July 2026.</p><h2>Security directors got the announcement. They didn’t get the playbook. </h2><p>“I’ve been in this industry for 27 years,” Cisco SVP and Chief Security and Trust Officer Anthony Grieco told VentureBeat in an exclusive interview at RSAC 2026. “I have never been more optimistic for what we can do to change security because of the velocity. It’s also a little bit terrifying because we’re moving so quickly. It’s also terrifying because our adversaries have this capability as well, and so frankly, we must move this quickly.”</p><p>Security directors saw this story told fifteen different ways this week, <a href="https://venturebeat.com/technology/anthropic-says-its-most-powerful-ai-cyber-model-is-too-dangerous-to-release">including VentureBeat’s exclusive interview with Anthropic’s Newton Cheng</a>.  As one widely shared X post summarizing the Mythos findings noted, the model cracked cryptography libraries, broke into a production virtual machine monitor, and gave engineers with zero security training working exploits by morning. What that coverage left unanswered: Where does the detection ceiling sit in the methods they already run, and what should they change before July?</p><h2>Seven vulnerability classes that show where every detection method hits its ceiling </h2><ol><li><p><a href="https://www.openbsd.org/"><b>OpenBSD</b></a><b> TCP SACK, 27 years old.</b> Two crafted packets crash any server. <a href="https://owasp.org/www-project-devsecops-guideline/latest/01b-Static-Application-Security-Testing">SAST</a>, fuzzers, and auditors missed a logic flaw requiring semantic reasoning about how TCP options interact under adversarial conditions. Campaign cost ~$20,000. Anthropic notes the $50 per-run figure reflects hindsight.</p></li><li><p><a href="https://code.ffmpeg.org/FFmpeg/FFmpeg/pulls/22499/files"><b>FFmpeg</b></a><b> H.264 codec, 16 years old.</b> Fuzzers exercised the vulnerable code path 5 million times without triggering the flaw, according to Anthropic.  Mythos caught it by reasoning about code semantics. Campaign cost ~$10,000.</p></li><li><p><a href="https://www.freebsd.org/"><b>FreeBSD</b></a><b> NFS remote code execution, </b><a href="https://red.anthropic.com/2026/mythos-preview/"><b>CVE-2026-4747</b></a><b>, 17 years old.</b> Unauthenticated root from the internet, per Anthropic’s assessment and independent reproduction. Mythos built a 20-gadget <a href="https://en.wikipedia.org/wiki/Return-oriented_programming">ROP chain</a> split across multiple packets. Fully autonomous.</p></li><li><p><b>Linux kernel local privilege escalation.</b> Mythos chained two to four low-severity vulnerabilities into full local privilege escalation via race conditions and <a href="https://en.wikipedia.org/wiki/Address_space_layout_randomization">KASLR</a> bypasses. <a href="https://cloudsecurityalliance.org/blog/2026/04/08/anthropic-s-mythos-is-here-defending-from-the-vulnpocalypse">CSA’s Rich Mogull noted</a> Mythos failed at remote kernel exploitation but succeeded locally. No automated tool chains vulnerabilities today.</p></li><li><p><b>Browser zero-days across every major browser.</b> Thousands identified. Some required human-model collaboration. In one case, Mythos chained four vulnerabilities into a <a href="https://en.wikipedia.org/wiki/JIT_spraying">JIT heap spray</a>, escaping both the renderer and the OS sandboxes. Firefox 147: 181 working exploits versus two for Opus 4.6.</p></li><li><p><b>Cryptography library vulnerabilities (</b><a href="https://en.wikipedia.org/wiki/Transport_Layer_Security"><b>TLS</b></a><b>, </b><a href="https://en.wikipedia.org/wiki/Galois/Counter_Mode"><b>AES-GCM</b></a><b>, </b><a href="https://en.wikipedia.org/wiki/Secure_Shell"><b>SSH</b></a><b>).</b> Implementation flaws enabling certificate forgery or decryption of encrypted communications, per <a href="https://red.anthropic.com/2026/mythos-preview/">Anthropic’s red team blog</a> and <a href="https://www.helpnetsecurity.com/2026/04/08/anthropic-claude-mythos-preview-identify-vulnerabilities/">Help Net Security</a>. A critical <a href="https://botan.randombit.net/">Botan</a> library certificate bypass was disclosed the same day as the Glasswing announcement. Bugs in the code that implements the math. Not attacks on the math itself.</p></li><li><p><a href="https://en.wikipedia.org/wiki/Hypervisor"><b>Virtual machine monitor</b></a><b> guest-to-host escape.</b> Guest-to-host memory corruption in a production VMM, the technology keeping cloud workloads from seeing each other’s data. Cloud security architectures assume workload isolation holds. This finding breaks that assumption.</p></li></ol><p>Nicholas Carlini, in Anthropic’s launch briefing: “I’ve found more bugs in the last couple of weeks than I found in the rest of my life combined.”</p><h2>VentureBeat&#x27;s prescriptive matrix </h2><table><tbody><tr><td><p>Vulnerability Class</p></td><td><p>Why Current Methods Miss It</p></td><td><p>What Mythos Does</p></td><td><p>Security Director Action</p></td></tr><tr><td><p><b>OS kernel logic</b> (OpenBSD 27yr, Linux 2-4 chain)</p></td><td><p>SAST lacks semantic reasoning. Fuzzers miss logic flaws. Pen testers time-boxed. Bounties scope-exclude kernel.</p></td><td><p>Chains 2-4 low-severity findings into local priv-esc. ~$20K campaign.</p></td><td><p>Add AI-assisted kernel review to pen test RFPs. Expand bounty scope. Request Glasswing findings from OS vendors before July. Re-score clustered findings by chainability.</p></td></tr><tr><td><p><b>Media codec</b> (FFmpeg 16yr H.264)</p></td><td><p>SAST unflagged. Fuzzers hit path 5M times, never triggered.</p></td><td><p>Reasons about semantics beyond brute-force. ~$10K campaign.</p></td><td><p>Inventory FFmpeg, libwebp, ImageMagick, libpng. Stop treating fuzz coverage as security proxy. Track Glasswing codec CVEs from July.</p></td></tr><tr><td><p><b>Network stack RCE</b> (FreeBSD 17yr, CVE-2026-4747)</p></td><td><p>DAST limited at protocol depth. Pen tests skip NFS.</p></td><td><p>Full autonomous chain to unauthenticated root. 20-gadget ROP chain.</p></td><td><p>Patch CVE-2026-4747 now. Inventory NFS/SMB/RPC services. Add protocol fuzzing to 2026 cycle.</p></td></tr><tr><td><p><b>Multi-vuln chaining</b> (2-4 sequenced, local)</p></td><td><p>No tool chains. Pen testers hours-limited. CVSS scores in isolation.</p></td><td><p>Autonomous local chaining via race conditions + KASLR bypass.</p></td><td><p>Require AI-assisted chaining in pen test methodology. Build chainability scoring. Budget AI red teams for 2026.</p></td></tr><tr><td><p><b>Browser zero-days</b> (thousands, 181 Firefox exploits)</p></td><td><p>Bounties + continuous fuzzing missed thousands. Some required human-model collaboration.</p></td><td><p>90x over Opus 4.6. Chained 4 vulns into JIT heap spray escaping renderer + OS sandbox.</p></td><td><p>Shorten patch SLA to 72hr critical. Pre-stage pipeline for July cycle. Pressure vendors for Glasswing timelines.</p></td></tr><tr><td><p><b>Crypto libraries</b> (TLS, AES-GCM, SSH, Botan bypass)</p></td><td><p>SAST limited on crypto logic. Pen testers rarely audit crypto depth. Formal verification not standard.</p></td><td><p>Found cert forgery + decryption flaws in battle-tested libraries.</p></td><td><p>Audit all crypto library versions now. Track Glasswing crypto CVEs from July. Accelerate PQC migration.</p></td></tr><tr><td><p><b>VMM / hypervisor</b> (guest-to-host memory corruption)</p></td><td><p>Cloud security assumes isolation. Few pen tests target hypervisor. Bounties rarely scope VMM.</p></td><td><p>Guest-to-host escape in production VMM.</p></td><td><p>Inventory hypervisor/VMM versions. Request Glasswing findings from cloud providers. Reassess multi-tenant isolation assumptions.</p></td></tr></tbody></table><h2>Attackers are faster. Defenders are patching once a year. </h2><p>The <a href="https://www.crowdstrike.com/en-us/global-threat-report/">CrowdStrike 2026 Global Threat Report</a> documents a 29-minute average eCrime breakout time, 65% faster than 2024, with an 89% year-over-year surge in AI-augmented attacks. CrowdStrike CTO Elia Zaitsev put the operational reality plainly in an exclusive interview with VentureBeat. “Adversaries leveraging agentic AI can perform those attacks at such a great speed that a traditional human process of look at alert, triage, investigate for 15 to 20 minutes, take an action an hour, a day, a week later, it’s insufficient,” Zaitsev said. A $20,000 Mythos discovery campaign that runs in hours replaces months of nation-state research effort.</p><p><a href="https://www.crowdstrike.com/en-us/">CrowdStrike</a> CEO George Kurtz reinforced that timeline pressure on <a href="https://www.linkedin.com/posts/georgekurtz_anthropic-claude-mythos-preview-the-more-activity-7447349706000826368-OG3E/">LinkedIn </a>the same day as the Glasswing announcement. &quot;AI is creating the largest security demand driver since enterprises moved to the cloud,&quot; Kurtz wrote. The regulatory clock compounds the operational one. The EU AI Act&#x27;s next enforcement phase takes effect August 2, 2026, imposing automated audit trails, cybersecurity requirements for every high-risk AI system, incident reporting obligations, and penalties up to 3% of global revenue. Security directors face a two-wave sequence: July&#x27;s Glasswing disclosure cycle, then August&#x27;s compliance deadline. </p><p>Mike Riemer, Field CISO at <a href="https://www.ivanti.com/">Ivanti</a> and a 25-year US Air Force veteran who works closely with federal cybersecurity agencies, told VentureBeat what he is hearing from the government. “Threat actors are reverse engineering patches, and the speed at which they’re doing it has been enhanced greatly by AI,” Riemer said. “They’re able to reverse engineer a patch within 72 hours. So if I release a patch and a customer doesn’t patch within 72 hours of that release, they’re open to exploit.” Riemer was blunt about where that leaves the industry. “They are so far in front of us as defenders,” he said.</p><p>Grieco confirmed the other side of that collision at RSAC 2026. “If you talk to an operational team and many of our customers, they’re only patching once a year,” Grieco told VentureBeat. “And frankly, even in the best of circumstances, that is not fast enough.”</p><p><a href="https://cloudsecurityalliance.org/blog/2026/04/08/anthropic-s-mythos-is-here-defending-from-the-vulnpocalypse">CSA’s Mogull</a> makes the structural case that defenders hold the long-term advantage: fix a vulnerability once and every deployment benefits. But the transition period, when attackers reverse-engineer patches in 72 hours and defenders patch once a year, favors offense.</p><p>Mythos is not the only model finding these bugs. Researchers at AISLE, an AI cybersecurity startup, <a href="https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier">tested Anthropic&#x27;s showcase vulnerabilities</a> on small, open-weights models and found that eight out of eight detected the FreeBSD exploit. AISLE says one model had only 3.6 billion parameters and costs 11 cents per million tokens, and that a 5.1-billion-parameter open model recovered the core analysis chain of the 27-year-old OpenBSD bug. AISLE&#x27;s conclusion: &quot;The moat in AI cybersecurity is the system, not the model.&quot; That makes the detection ceiling a structural problem, not a Mythos-specific one. Cheap models find the same bugs. The July timeline gets shorter, not longer. </p><p>Over 99% of the vulnerabilities Mythos has identified have not yet been patched, per Anthropic’s red team blog. The public Glasswing report lands in early July 2026. It will trigger a high-volume patch cycle across operating systems, browsers, cryptography libraries, and major infrastructure software. Security directors who have not expanded their patch pipeline, re-scoped their bug bounty programs, and built chainability scoring by then will absorb that wave cold. July is not a disclosure event. It is a patch tsunami.</p><h2>What to tell the board </h2><p>Every security director tells the board “we have scanned everything.” Merritt Baer, CSO at <a href="https://www.enkryptai.com/">Enkrypt AI</a> and former Deputy CISO at AWS, told VentureBeat that the statement does not survive Mythos without a qualifier.</p><p>“What security leaders actually mean is: we have exhaustively scanned for what our tools know how to see,” Baer said in an exclusive interview with VentureBeat. “That’s a very different claim.”</p><p>Baer proposed reframing residual risk for boards around three tiers: known-knowns (vulnerability classes your stack reliably detects), known-unknowns (classes you know exist but your tools only partially cover, like stateful logic flaws and auth boundary confusion), and unknown-unknowns (vulnerabilities that emerge from composition, how safe components interact in unsafe ways). “This is where Mythos is landing,” Baer said.</p><p>The board-level statement Baer recommends: “We have high confidence in detecting discrete, known vulnerability classes. Our residual risk is concentrated in cross-function, multi-step, and compositional flaws that evade single-point scanners. We are actively investing in capabilities that raise that detection ceiling.”</p><p>On chainability, Baer was equally direct. “Chainability has to become a first-class scoring dimension,” she said. “CVSS was built to score atomic vulnerabilities. Mythos is exposing that risk is increasingly graph-shaped, not point-in-time.” Baer outlined three shifts security programs need to make: from severity scoring to exploitability pathways, from vulnerability lists to vulnerability graphs that model relationships across identity, data flow, and permissions, and from remediation SLAs to path disruption, where fixing any node that breaks the chain gets priority over fixing the highest individual CVSS.</p><p>“Mythos isn’t just finding missed bugs,” Baer said. “It’s invalidating the assumption that vulnerabilities are independent. Security programs that don’t adapt, from coverage thinking to interaction thinking, will keep reporting green dashboards while sitting on red attack paths.”</p><p><i>VentureBeat will update this story with additional operational details from Glasswing&#x27;s founding partners as interviews are completed.</i></p>]]></description>
            <author>louiswcolumbus@gmail.com (Louis Columbus)</author>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/6U30KdpUttFqCNrnyaxxpJ/35987bc5f4459240fe7195e361f84706/Anthropic_Mythos_Hero.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Claude, OpenClaw and the new reality: AI agents are here — and so is the chaos]]></title>
            <link>https://venturebeat.com/infrastructure/claude-openclaw-and-the-new-reality-ai-agents-are-here-and-so-is-the-chaos</link>
            <guid isPermaLink="false">E4kZwK085N3OHabqiT6mh</guid>
            <pubDate>Wed, 08 Apr 2026 22:26:37 GMT</pubDate>
            <description><![CDATA[<p>The age of agentic AI is upon us — whether we like it or not. What started with an innocent question-answer banter with ChatGPT back in 2022 has become an existential debate on job security and the rise of the machines. </p><p>More recently, fears of reaching artificial general intelligence (AGI) have become more real with the advent of  powerful autonomous agents like Claude Cowork and <a href="https://venturebeat.com/security/openclaw-500000-instances-no-enterprise-kill-switch">OpenClaw</a>. Having played with these tools for some time, here is a comparison.</p><p>First, we have OpenClaw (formerly known as Moltbot and Clawdbot). Surpassing 150,000 GitHub stars in days, OpenClaw is already being deployed on local machines with deep system access. This is like a robot “maid” (Irona for <i>Richie Rich</i> fans, for instance) that you give the keys to your house. It’s supposed to clean it, and you give it the necessary autonomy to take actions and manage your belongings (files and data) as it pleases. The whole purpose is to perform the task at hand — inbox triaging, auto-replies, content curation, travel planning, and more.</p><p>Next we have Google’s <a href="https://antigravity.google/">Antigravity</a>, a coding agent with an IDE that accelerates the path from prompt to production. You can interactively create complete application projects and modify specific details over individual prompts. This is like having a junior developer that can not only code, but build, test, integrate, and fix issues. In the realworld, this is like hiring an electrician: They are really good at a specific job and you only need to give them access to a specific item (your electric junction box). </p><p>Finally, we have the mighty Claude. The release of Anthropic&#x27;s Cowork, which featured AI agents for automating legal tasks like contract review and NDA triage, caused a sharp sell-off in legal-tech and software-as-a-service (SaaS) stocks (referred to as the <a href="https://venturebeat.com/infrastructure/intuit-is-betting-its-40-years-of-small-business-data-can-outlast-the">SaaSpocalypse</a>). Claude has anyway been the go-to chatbot; now with Cowork, it has domain knowledge for specific industries like legal and finance. This is like hiring an accountant. They know the domain inside-out and can complete taxes and manage invoices. Users provide specific access to highly-sensitive financial details.</p><h2>Making these tools work for you</h2><p>The key to making these tools more impactful is giving them more power, but that increases the <a href="https://venturebeat.com/security/openclaw-can-bypass-your-edr-dlp-and-iam-without-triggering-a-single-alert">risk of misuse</a>. Users must trust providers like Anthorpic and Google to ensure that agent prompts will not cause harm, leak data, or provide unfair (illegal) advantage to certain vendors. OpenClaw is open-source, which complicates things, as there is no central governing authority. </p><p>While these technological advancements are amazing and meant for the greater good, all it takes is one or two adverse events to cause panic. Imagine the agentic electrician frying all your house circuits by connecting the wrong wire. In an agent scenario, this could be injecting incorrect code, breaking down a bigger system or adding hidden flaws that may not be immediately evident. Cowork could miss major saving opportunities when doing a user&#x27;s taxes; on the flip side, it could include illegal writeoffs. Claude can do unimaginable damage when it has more control and authority.</p><p>But in the middle of this chaos, there is an opportunity to really take advantage. With the right guardrails in place, agents can focus on specific actions and avoid making random, unaccounted-for decisions. Principles of responsible AI — accountability, transparency, reproducibility, security, privacy — are extremely important. Logging agent steps and human confirmation are absolutely critical. </p><p>Also, when agents deal with so many diverse systems, it&#x27;s important they speak the same language. Ontology becomes very important so that events can be tracked, monitored, and accounted for. A shared domain-specific ontology can define a “code of conduct.&quot; These ethics can help control the chaos. When tied together with a shared trust and distributed identity framework, we can build systems that enable agents to do truly useful work. </p><p>When done right, an agentic ecosystem can greatly offload the human “cognitive load” and enable our workforce to perform high-value tasks. Humans will benefit when agents handle the mundane.</p><p><i>Dattaraj Rao is innovation and R&amp;D architect at Persistent Systems. </i></p>]]></description>
            <author>dattarajraogravitar@gmail.com (Dattaraj Rao, Persistent Systems)</author>
            <category>Infrastructure</category>
            <category>DataDecisionMakers</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/33OB5cKXtts9VZ7sMyzGew/7454f3b529fbde6e78746d28b720e4c4/Chaos.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
    </channel>
</rss>