RL Blog

Topics

All Blog PostsAppSec & Supply Chain SecurityDev & DevSecOpsProducts & TechnologySecurity OperationsThreat Research
Why RL Built Spectra Assure Community
April 14, 2026

Why RL Built Spectra Assure Community

We set out to help dev and AppSec teams secure the village: OSS dependencies, malware, more. Learn how.

Read More about Why RL Built Spectra Assure Community
Why RL Built Spectra Assure Community

Follow us

XX / TwitterLinkedInLinkedInFacebookFacebookInstagramInstagramYouTubeYouTubeblueskyBluesky

Subscribe

Get the best of RL Blog delivered to your in-box weekly. Stay up to date on key trends, analysis and best practices across threat intelligence and software supply chain security.

ReversingLabs: The More Powerful, Cost-Effective Alternative to VirusTotalSee Why
Skip to main content
Contact UsSupportLoginBlogCommunity
reversinglabsReversingLabs: Home
Solutions
Secure Software OnboardingSecure Build & ReleaseProtect Virtual MachinesIntegrate Safe Open SourceGo Beyond the SBOM
Increase Email Threat ResilienceDetect Malware in File Shares & StorageAdvanced Malware Analysis SuiteICAP Enabled Solutions
Scalable File AnalysisHigh-Fidelity Threat IntelligenceCurated Ransomware FeedAutomate Malware Analysis Workflows
Products & Technology
Spectra Assure®Software Supply Chain SecuritySpectra DetectHigh-Speed, High-Volume, Large File AnalysisSpectra AnalyzeIn-Depth Malware Analysis & Hunting for the SOCSpectra IntelligenceAuthoritative Reputation Data & Intelligence
Spectra CoreIntegrations
Industry
Energy & UtilitiesFinanceHealthcareHigh TechPublic Sector
Partners
Become a PartnerValue-Added PartnersTechnology PartnersMarketplacesOEM Partners
Alliances
Resources
BlogContent LibraryCybersecurity GlossaryConversingLabs PodcastEvents & WebinarsLearning with ReversingLabsWeekly Insights Newsletter
Customer StoriesDemo VideosDocumentationOpenSource YARA Rules
Company
About UsLeadershipCareersSeries B Investment
EventsRL at RSAC
Press ReleasesIn the News
Pricing
Software Supply Chain SecurityMalware Analysis and Threat Hunting
Request a demo
Menu
AppSec & Supply Chain SecurityMay 13, 2026

Think AI agents are risky? Try the underlying stack

To manage risk from agentic AI, organizations need to focus on the infrastructure they run on.

man in suit
Jaikumar Vijayan, Freelance technology journalistJaikumar Vijayan
FacebookFacebookXX / TwitterLinkedInLinkedInblueskyBlueskyEmail Us
AI infrastructure

Organizations rushing to deploy AI agents across their workflows need to realize that the underlying agent stack that governs identity, access, persistence, security, and operational control is more consequential than the particular model chosen. 

The engineering leader and director for Google Cloud AI, Addy Osmani, wrote in a recent Substack post that in many environments, agents are being granted broad autonomy without the infrastructure needed to properly control, monitor, or audit their behavior. As those systems become more deeply embedded in enterprise operations, weaknesses in the underlying stack can metastasize from being an engineering concern to become a broader security, governance, and operational risk.

"Right now, the industry is living with what I’d call excessive agency: autonomous systems given broad permissions to get things done, then left to discover — at runtime, in production — that a schema drifted, an API changed, or a downstream service started returning PII it wasn’t supposed to. This is not a failure of the people building agents. It is a failure of the stack they’re building on.”
—Addy Osmani

Here's why agentic AI infrastructure security is critical to managing risk — and four essential requirements.

[ See webinar: AI Redefines Software Risk: Develop a New Playbook ]

Agents are everywhere now

Osmani’s concern is one of the current moment. A growing number of organizations are feverishly deploying AI agents, automating everything from triaging support tickets to generating code, responding to customer inquiries, updating account records, and handling procurement workflows and invoice processing.

A recent survey by Salesforce found that 82% of IT leaders are either using AI agents at their organizations or planning to deploy them within the next two years. Many are already running pilots across HR, IT, finance, and customer service workflows. Similarly, McKinsey’s 2025 State of AI report found that 62% of organizations are either piloting or using AI agents to automate key aspects of their business and operations.

McKinsey reported that organizations most commonly are using agents to manage service desks and do deep research. It added that organizations in the technology, media, telecommunications, and health care sectors are using agents more than those in  other sectors.

Rethinking the agent stack

Osmani said effective AI agent systems require a fundamental rethink of the stack on which they are deployed. He listed four requirements:

  • Context: Agents need broad, unified access to enterprise context rather than fragmented data so they can reason better.
  • Identity: They must have their own verifiable identities so their actions can be attributed and controlled.
  • Preservation: For meaningful enterprise work, such as compliance reviews, procurement processes, or incident response investigations that can span days and weeks, agents need to be able to reliably preserve state and resume tasks from where they left off, after interruptions. 
  • Standardization: Organizations must standardize their platforms so they’re not repeatedly rebuilding the same infrastructure for memory, observability, orchestration, and tooling integration.

Ensar Seker, CISO at SOCRadar, agrees with Osmani’s assessment. “The biggest risk with AI agents is not only the model; it is the architecture around the model,” he said. 

“Many organizations are building agents on top of fragile plumbing: shared service accounts, custom session logic, weak audit trails, and inconsistent access controls.”
—Ensar Seker

They are creating a security problem before the agent even makes its first decision, Seker said.

The core failure point

With AI agent deployments, Seker said, the core failure point is usually identity and permissions design. A common mistake is to treat agents like scripts or automations when they actually behave more like digital workers. If your agents don’t have clear identities, scoped permissions, logging, and revocation paths, he said, you can lose control over what the agent did, why it did it, and which data it touched.

David Brumley, chief AI and science officer at Bugcrowd, said that giving every agent its own identity ensures better auditability, accountability, and access control.

“You would never ask people to share one password and one account, but that is effectively what a lot of organizations are asking agents to do today.”
—David Brumley

Agent identities give organizations a way to scope permissions, trace actions, and revoke access when needed, he said.

Putting things in context

In another echo of Osmani’s requirements, Eric Hulse, director of research at Command Zero, said context is more important than most organizations realize. Many agents have access to only a narrow slice of enterprise information at any given moment, but real enterprise work requires reasoning across CRM platforms, ERP systems, ticketing tools, project plans, transcripts, analytics systems, data warehouses, and other systems. “If an agent investigates an alert with access to endpoint telemetry but no visibility into network traffic, cloud logs, or identity context, it’s not going to say, ‘I don’t have enough information.’ It’s going to synthesize a verdict from what it has,” Hulse said.

“A confident wrong answer looks identical to a confident right answer.”
—Eric Hulse

Bugcrowd’s Brumley stressed: “Agents often don’t know what they don’t know.”

Real investigation means following a thread wherever it goes, Hulse said: from an endpoint artifact into the identity layer, into cloud API logs, into network egress. “That requires unified context at the platform level. Not stitched-together API calls at runtime that happen to return data in the right order on a good day.”

But organizations need to be cautious about how they enable context for agents, cautioned Brumley. More context is not always better or safer, he said. The best approach is to give agents limited, purpose-built access to what they need for a specific task — bounded context, not universal context. 

Why agents need to be persistent

Osmani said agents also need to be able to maintain continuity. Most AI agents today can handle short, contained tasks but aren’t up to what genuine enterprise deployment demands. Procurement cycles, compliance audits, incident investigations, and other real business workflows can span days, weeks, or even months, and having agents that simply stop when they hit a time or processing limits forces humans to pick up from where the agent left off. 

What enterprises need are agents that can maintain progress across interruptions, hand off work without losing context, and leave a clear, auditable trail, while still being intelligent enough to know when to pause and ask for human guidance.

“Enterprise-grade autonomy requires durable, cloud-native execution with a much higher floor than ‘The session stayed up.’”
—Addy Osmani

How to get started on agentic AI stack security

When architecting an agent stack, start with a clear idea of what you’re building for, Hulse said.  Most agentic security architectures give alerts and resolutions the one-for-one treatment: one alert in, one verdict out. That’s the wrong model. “Security work is investigation-shaped,” Hulse said.

“You start with a signal, follow threads, accumulate evidence across multiple systems and time windows, and arrive at a conclusion that might span several alerts that, individually, looked unrelated. If your architecture forces every alert to be a self-contained task, you’ve inherited every limitation of the triage layer and put an AI label on it.”
—Eric Hulse

Before building in autonomy, focus on identity and context, Hulse said, because without them, governance debt will inevitably compound. “You will hit an incident that exposes the shared credentials, the missing audit trail, the agent that had way too much access for what it was supposed to do.”

And before planning to scale, measure accuracy, not throughput. That means doing things like tracking false-positive and false-negative rates by alert type, data source, and time of day, Hulse said.

“The organizations that do this end up with agents that get measurably better over time because the feedback loop is short and honest. The ones chasing closure rates end up with fast agents that are confidently wrong at scale — worse than the manual process they replaced.”

Keep learning

  • Get up to speed on the state of software security with RL's Software Supply Chain Security Report 2026. Plus: See the the webinar to discussing the findings.
  • Learn why binary analysis is a must-have in the Gartner® CISO Playbook for Commercial Software Supply Chain Security.
  • Take action on securing AI/ML with our report: AI Is the Supply Chain. Plus: See RL's research on nullifAI and watch how RL discovered the novel threat.
  • Get the report: Go Beyond the SBOM. Plus: See the CycloneDX xBOM webinar.

Explore RL's Spectra suite: Spectra Assure for software supply chain security, Spectra Detect for scalable file analysis, Spectra Analyze for malware analysis and threat hunting, and Spectra Intelligence for reputation data and intelligence.

Tags:AppSec & Supply Chain SecurityArtificial Intelligence (AI)/Machine Learning (ML)

More Blog Posts

Shai-hulud worm DevOps

Mini Shai-Hulud tears at OSS trust

This latest Team PCP compromises of popular and foundational npm packages are rocking open source. 

Learn More about Mini Shai-Hulud tears at OSS trust
Mini Shai-Hulud tears at OSS trust
NVD enrichment

Selective NVD enrichment: Why it matters

AI vulnerability reporting is overwhelming teams — and NIST. But for AppSec, scaling back analysis is cause for alarm.

Learn More about Selective NVD enrichment: Why it matters
Selective NVD enrichment: Why it matters
math strategy

How Mythos changes the AppSec calculus

Here are the facts on Claude Mythos — and why a layered application security framework is essential.

Learn More about How Mythos changes the AppSec calculus
How Mythos changes the AppSec calculus
Trust model flips

How agentic AI flips the trust model

As AppSec shifts focus from the components to data, your strategy needs updating. Are you on top of your trust debt?

Learn More about How agentic AI flips the trust model
How agentic AI flips the trust model

Spectra Assure Free Trial

Get your 14-day free trial of Spectra Assure for Software Supply Chain Security

Get Free TrialMore about Spectra Assure Free Trial
Blog
Events
About Us
Webinars
In the News
Careers
Demo Videos
Cybersecurity Glossary
Contact Us
reversinglabsReversingLabs: Home
Privacy PolicyCookiesImpressum
All rights reserved ReversingLabs © 2026
XX / TwitterLinkedInLinkedInFacebookFacebookInstagramInstagramYouTubeYouTubeblueskyBlueskyRSSRSS
Back to Top