Building the Internet of Agents: A Trust Layer for the Next Web

Insights from Vijoy Pandey, Cisco Outshift, and the Confidential Summit

“A human can’t do much damage in an hour.
An agent acting like a human—at machine speed—can do a lot.”
– Vijoy Pandey, SVP & GM, Cisco

We’re entering the era of agentic AI: networks of autonomous, collaborative agents that behave like humans but act at machine speed and scale. They build, decide, communicate, and self-replicate. But there’s one thing they can’t yet do—earn our trust.

At the Confidential Summit two weeks ago in San Francisco, that challenge took center stage. Executives and builders from NVIDIA, Microsoft Azure, Google Cloud, AWS, Intel, ARM, AMD, ServiceNow, LangChain, Anthropic, DeepMind, and more came together to ask a hard question:

Can we build an Internet of Agents that is open, interoperable—and trusted?

The answer is yes! And many came prepared with reference architectures, including OPAQUE.

In this episode of AI Confidential, we sat down with Vijoy Pandey, who leads Cisco’s internal incubator Outshift and the industry initiative Agency. Along with co-host Mark Hinkle, we explored why this problem can’t be solved with policy patches or paper governance.

🧠 From Deterministic APIs to Probabilistic Agents

Today’s internet runs on deterministic computing—you know what API you’re hitting and what result to expect. Agents break that model.

Agentic systems introduce probabilistic logic, dynamic behavior, and autonomous decision-making. One input can lead to many outcomes. That’s powerful—but also dangerous.

🔐 Why We Need a Trust Layer

As Vijoy put it: “We’ve built access control lists, compliance programs, and identity providers—for humans. None of those scale to agentic systems.”

Agents can impersonate employees, leak IP, or introduce bias—without ever breaking a rule on paper. That’s why verifiable trust is the new foundation.

At the Confidential Summit, dozens of companies showcased confidential AI stacks that create cryptographic guarantees at runtime—across data, identity, code, and communication.

🌐 Introducing the Internet of Agents

The future isn’t a single AI. It’s collaborative networks of agents, working across clouds, enterprises, and toolchains. Vijoy’s team at Agency (agency.org) is building the open-source fabric for this new internet: discoverable, composable, verifiable agents that speak a shared language.

OPAQUE has joined this effort to help embed verifiable, hardware-enforced trust into the open stack. And others—from LangChain to Galileo, Cisco to CrewAI—are building multi-agent systems for real enterprise workflows.

🚀 Use Cases Are Here

This isn’t science fiction. ServiceNow is already using OPAQUE-powered confidential agents to accelerate sales operations. Cisco’s SRE teams have offloaded 30% of their infrastructure workflows to Jarvis, a composite agent framework with 20+ agents and 50+ tools.

These are just the beginning.

🧱 A Call to Architects

The trust layer of the Internet of Agents is being designed right now—at the protocol layer, at the hardware layer, and in the open. It will require open standards, decentralized identity, hardware attestation, and zero-trust workflows by default.

The risks are massive. The opportunity is bigger. But trust can’t be retrofitted. It has to be built in.

Listen to the full convesaration with Vijoy Pandey –> Spotify Apple Podcast YouTube

And you can find all our podcast episodes –> https://podcast.aiconfidential.com and you can subscribe to our newsletter –> https://aiconfidential.com

Confidential Summit Wrap

We just wrapped the Confidential Summit in SF—and it was electric.
From NVIDIA, Arm, AMD, and Intel Corporation to Microsoft, Google and Anthropic the world’s leading builders came together to answer one critical question:

**How do we build a verifiable trust layer for AI and the Internet?**

🔐 Ion Stoica (SkyLab/Databricks) reminded us: as agentic systems scale linearly, risk compounds exponentially.

🧠 Jason Clinton (Anthropic) stunned with stats:
→ 65% of Anthropic’s code is written by Claude. By year’s end? 90–95%.
→ AI compute needs are growing 4x every 12 months.
→ “This is the year of the agent,” he said—soon we’ll look back on it like we do Gopher.

🛠️ Across the board, Big Tech brought reference architectures for Confidential AI:

→Microsoft shared real-world Confidential AI infrastructure running in Azure
→Meta detailed how WhatsApp uses Private Processing to secure messages
→Google, Apple, and TikTok revealed their confidential compute strategies
→OPAQUE launched a Confidential Agent stack built on NVIDIA NeMo + LangGraph with verifiable guarantees before, during, and after agent execution
→ AMD also had exciting new confidential product announcements.

🎯 But here’s the real takeaway:
– This wasn’t a vendor expo. It was a community and ecosystem summit, a collaboration that culminated in a shared commitment.
– Over the next 12 months, leaders from Google, Microsoft, Anthropic, Accenture, AMD, Intel, NVIDIA, and others will collaborate to release a reference architecture for an open, interoperable Confidential AI stack. Think Confidential MCP with verifiable guarantees.

We’re united in building a trust layer for the agentic web. And it’s going to take an ecosystem and community. What we build now—with this ecosystem, this community—will shape how the world relates to technology for the next century. And more importantly, how we relate to each other, human to human.

Subscribe to AIConfidential.com to get the sessions, PPTs, videos, and podcast drops.

Thank you to everyone who joined us—on site, remote, or behind the scenes. Let’s keep building to ensure AI can be harnessed to advance human progress.

AI at the Edge: Governance, Trust, and the Data Exhaust Problem

What enterprises must learn—from history and from hackers—to survive the AI wave

“The first thing I tell my clients is: Are you accepting that you’re getting probabilistic answers? If the answer is no, then you cannot use AI for this.”
— John Willis, enterprise AI strategist

AI isn’t just code anymore. It’s decision-making infrastructure. And in a world where agents can operate at machine speed, acting autonomously across systems and clouds, we’re encountering new risks—and repeating old mistakes.

In this episode of AI Confidential, we’re joined by industry legend John Willis, who brings four decades of experience in operations, devops, and AI strategy. He’s the author of The Rebels of Reason, a historical journey through the untold stories of AI’s pioneers—and a stark warning to today’s enterprise leaders.

Here are the key takeaways from our conversation:

🔄 History Repeats Itself—Unless You Design for It

John’s central insight? Enterprise IT keeps making the same mistakes. Shadow IT, ungoverned infrastructure, and tool sprawl defined the early cloud era—and they’re back again in the age of GenAI. “We wake up from hibernation, look at what’s happening, and say: what did y’all do now?”

🤖 AI is Probabilistic—Do You Accept That?

Too many leaders expect deterministic behavior from fundamentally probabilistic systems. “If you’re building a high-consequence application, and you’re not accepting that LLMs give probabilistic answers, you’re setting yourself up to fail,” John warns.

This demands new tooling, new culture, and new operational rigor—including AI evaluation pipelines, attestation mechanisms, and AI-specific gateways.

📉 The Data Exhaust is Dangerous

Data isn’t just an input—it’s an output. And that data exhaust can now be weaponized. Whether it’s customer interactions, supply chain patterns, or software development workflows, LLMs are remarkably good at inferring proprietary IP from metadata alone.

“Your cloud provider—or their contractor—could rebuild your product from the data exhaust you’re streaming through their APIs,” John notes. If you’re not using attested, verifiable systems to constrain where and how your data flows, you’re building your own future competitor.

🛡️ Governance, Attestation, and Confidential AI

Confidential computing may sound like hardware tech, but its real value lies in guarantees: provable, cryptographic enforcement of data privacy and policy at runtime.

OPAQUE’s confidential AI fabric is one example—enabling encrypted data pipelines, agentic policy enforcement, and hardware-attested audit trails that align with enterprise governance requirements. “I didn’t care about the hardware,” John admits. “But once I saw the guarantees you get, I was all in.”

📚 Why the History of AI Still Matters

John’s latest book, The Rebels of Reason, brings to life the hidden history of AI—spotlighting unsung pioneers like Fei-Fei Li and Grace Hopper. “Without ImageNet, we don’t get AlexNet. Without Hopper’s compiler, we don’t get natural language programming,” he explains.

Understanding AI’s history isn’t nostalgia—it’s necessary context for navigating where we’re going next. Especially as we transition into agentic systems with layered, distributed, and dynamic behavior.


If you’re an enterprise CIO, CISO, or builder, this episode is your field guide to what’s coming—and how to avoid becoming the next cautionary tale.

Listen to the full episode here: Spotify | Apple Podcast | YouTube

And you can find all our podcast episodes –> https://podcast.aiconfidential.com, and you can subscribe to our newsletter –> https://aiconfidential.com