The conversation has changed from what AI can do to what it takes to run it. That shift is driving unprecedented focus on the foundational layers of IT and the AI Networking Summit cover it all: Infrastructure, Networking, Automation, and Security. The AI Networking Summit brings these domains together into four dedicated tracks designed to help enterprises move from AI pilots to scalable, secure deployment. This is where strategy meets implementation and where the AI-ready enterprise takes shape.
Build the foundation for enterprise AI. This track covers private infrastructure, cost models, and scaling AI workloads in production.
We know most Enterprises aren’t running tens of thousands of GPUs (yet!), but AI is quickly accelerating every industry. Before you can run AI, you need a network built to carry it. That means your network must be smart, secure, and ready to evolve. This session cuts through the complexity with a practical look at how enterprises can get AI infrastructure up and running without disrupting existing workloads. Learn about what it takes to run both traditional and AI workloads seamlessly, with unified operations, automation, and real-time observability across traditional and AI fabrics. Your blueprint to success is a proven modernization path. No matter where you are on your AI journey, you’ll come away with practical ways to simplify, future-proof, and supercharge your data center for the Agentic Era.
AI has evolved from passive insights to active decision-making—drafting emails, generating code, and executing tasks across enterprise systems. As AI agents grow more powerful, they also introduce unpredictability, expanding attack surfaces and compliance risks. In this session, Netskope shows how to move beyond simply enabling AI to securing it at every touchpoint. We’ll outline a practical blueprint to protect AI agents end-to-end across users, applications, and data.
Learn how to maintain visibility into agent behavior, apply advanced DLP and threat protection to reduce misuse, and build an AI security strategy that balances innovation with governance.
Across the Global 2000, most AI initiatives remain stuck in experimentation and proof-of-concept phases. The reasons are clear: GPU and accelerator costs remain high, hardware obsolescence cycles are accelerating, skilled infrastructure talent is scarce, equipment lead times are unpredictable, and many enterprises lack the power, cooling, and space required to scale AI infrastructure.
This session examines how the vendor ecosystem is responding with new silicon, system architectures, networking fabrics, managed AI infrastructure, and consumption models designed to reduce risk and accelerate enterprise adoption. The discussion will focus on how enterprises can move beyond experimentation to deploy production AI infrastructure that enables true business transformation and competitive advantage.
Key Questions
Takeaways
Redesign the network for the AI era with agentic overlays, A2A fabrics, and architectures built for autonomous systems.
As AI models continue to scale, both training and inference are growing rapidly in operational importance. Training pushes the limits of compute density and interconnect scale, while inference now dominates production workloads. Together, these forces are reshaping AI system architectures.
Meeting these demands requires a next-generation networking fabric that can:
We will present the latest advancements in industry initiatives—including Ethernet Scale-Up Networking (ESUN), Scale-Up Ethernet Transport (SUE-T), and Open Cluster Design for AI—and show how Ethernet is democratizing large-scale AI deployments through insights from G42 and other AI operators.
Enterprise network and security operations centers are sitting on a goldmine of untapped efficiency — not from bleeding-edge autonomous agentic AI, but from mature, battle-tested capabilities already at their fingertips. This panel brings together practitioners who have deployed large language models, machine learning pipelines, and intelligent automation within live NOC and SOC environments to cut mean-time-to-resolution, reduce alert fatigue, and reclaim analyst capacity. Panelists will share concrete before-and-after metrics, walk through implementation playbooks that worked — and the integration pitfalls, data-quality traps, and organizational resistance that nearly derailed them. If your team is still triaging thousands of alerts manually or troubleshooting network issues with legacy runbooks, this session delivers the pragmatic roadmap to operationalize AI capabilities that are proven, available today, and delivering measurable ROI.
Agentic AI overlays promise networks that sense, reason, and act—but the real challenge is treating AI agents as first-class identities in the network. Building on ONUG’s A2A and Agentic Overlay concepts, this session explores how to standardize identity, trust, and messaging among agents, and how those agents interact with the underlying network fabric.
Key Questions: – What should an A2A reference architecture look like in a large enterprise (broker vs. mesh, policy dialects, schemas)? – How do we provide Zero-Trust identity and least-privilege capability routing for agents issuing network changes? – What telemetry is needed to link agent intents → network changes → workload outcomes → cost? – How do you avoid “agent sprawl” and conflicting policies across multiple vendors’ AI assistants?
Takeaways: – A clear mental model of Agentic Overlays and A2A in networking. – Governance patterns for letting agents safely orchestrate NaaS, routing, and provisioning.
As traffic demands increased, eBay needed to scale network bandwidth without sacrificing cost efficiency or operational predictability. In this session, eBay will discuss how it used open networking and SONiC to move from 100G to 400G, enabling faster upgrades, greater hardware flexibility, and stronger long-term operational control. Attendees will gain practical insight into how standardized, vendor-agnostic networking can help modernize infrastructure while supporting scale, efficiency, and consistent NetOps operations.
As campus networks transition from general connectivity to AI-ready infrastructure, the surge in LLM inference traffic demands a fundamental architectural shift. Localized AI applications require high throughput and ultra-low latency that legacy systems weren’t built to handle. This session explores how SONiC’s open-source framework provides the telemetry, load balancing, and agility needed to adapt campus fabrics for heavy inference workloads. Join us for a deep dive into the foundational capabilities of SONiC that turn bandwidth-heavy AI demands into seamless learning experiences.
ISP incidents remain one of the fastest ways to impact user experience and one of the hardest to prove quickly. In this session, we’ll walk through a real customer example and show how we detected an ISP issue early by comparing user-to-app path performance across regions and ISPs. We’ll also cover how “peer” telemetry (crowd signals across many environments) can help you validate whether an issue is inside your network, at an ISP, or specific to a single site.
We’ll focus on the practical workflow: spotting the anomaly, confirming it’s an ISP event, narrowing blast radius, and taking action—like rerouting users to a healthier egress/data center—to cut troubleshooting time dramatically.
Networking plays a critical role in the accelerated compute fabric. Critical features to enable high performance networking for AI workloads are being defined and ratified in UEC and ESUN (OCP). Marvell is a contributor to both these industry standards and has developed a full portfolio of Scale-out and Scale-up switches powered by SONiC to enable best in class AI fabrics.
eBay operates one of the most advanced enterprise networks built on SONiC and large-scale automation. In this technical deep dive, Rick Casarez provides an inside look at the architecture, design principles, and operational decisions behind eBay’s production SONiC deployment.
Rick will also demonstrate how eBay is applying AI-powered LLMs to assist and automate NOC workflows, enabling engineers to investigate incidents faster, interpret telemetry, and streamline troubleshooting runbooks. The session concludes with a look at how eBay’s network architecture is evolving to support AI workloads and more autonomous network operations.
Key Topics Covered
• Architecture of eBay’s production SONiC-based network
• Design principles behind eBay’s disaggregated networking strategy
• Hardware abstraction and multi-vendor switching architecture
• Automation frameworks and infrastructure-as-code for network operations
• Telemetry, observability, and operational tooling at scale
• Applying LLMs to assist and automate NOC workflows
• AI-assisted troubleshooting, incident investigation, and operational runbooks
• Preparing enterprise networks for AI infrastructure and high-performance workloads
Who Should Attend
• Enterprise network engineers and architects evaluating or deploying SONiC
• NOC and operations leaders exploring AI-assisted operations
• Infrastructure engineers responsible for automation and network reliability
• Platform and SRE teams operating large-scale distributed infrastructure
• IT leaders preparing networks to support AI and high-performance compute environments
As enterprise networks become more critical — and more complex — the limits of traditional network performance monitoring are increasingly exposed. In today’s AI-driven digital environment, networks remain one of the most static and risk-averse layers of infrastructure, even as business demands for agility, resilience, and real-time insight continue to rise.
Join IDC’s Mark Leary for a data-driven look at how the bar for network observability has been raised — and what IT leaders and suppliers must do to meet the moment
For most organizations, the low-hanging network issues are solved; thresholds, polling, and playbooks handle them. The remainder? Issues that cross domain boundaries, evade static rules, and cannibalize NOC time while defying existing solutions. At IBM, we believe the right AI architecture can transform this space: time-series foundation models that observe without thresholds, paired with agentic reasoning that forms root cause hypotheses across your full stack. This session covers why the architecture matters and how context engineering, composable skills, and the open Model Context Protocol (MCP) create an extensible AI pipeline enterprises can localize and govern, with human oversight at every decision point. You’ll leave with a clear framework for evaluating where AI fits in your operations model and what it takes to build trust between NOC teams and intelligent systems.
Agentic Networking is the next evolution of cloud and network operations, where intelligent, autonomous agents – powered by LLMs, real-time telemetry, and policy-driven intent – continuously observe, decide, and act across complex, multi-cloud environments. Instead of manually tuning networks or relying solely on static automation, agentic systems interpret high-level business intents (minimize latency for this workload, shift traffic for cost efficiency, isolate anomalies before impact), translate them into actionable network changes, and coordinate with other agents to optimize end-to-end performance. This creates a self-healing, self-optimizing fabric that adapts dynamically to demand, failures, and emerging threats.
We will discuss how Agentic Networking transforms the network from a static transport layer into an intelligent, collaborative control plane: accelerating operations, reducing risk, and enabling entirely new levels of agility for AI-era workloads.
Building on the NYC “From NOC to AOC” concept, this session focuses on operational trust: how to evolve from human-centric operations centers to Agentic Operations Centers (AOCs) where AI agents handle most monitoring and remediation—but humans still own accountability. It ties directly into ONUG’s 2026 focus on agentic operations you can trust, including standard identity, guardrails, and observability.
Key Questions: – What work should be fully handed to agents in 2026 (Tier-1, Tier-2 incidents, routine changes), and what must remain human-centric? – How do you train and certify “agent supervisors” to oversee closed-loop systems without being buried in noise? – How do you design human-in-the-loop patterns that are fast enough for AI-era incidents yet still satisfy governance? – What changes in org design, skills, and incentives are needed for teams to trust automation?
Takeaways: – A target operating model for AOCs, including roles, skills, and escalation patterns. – Examples of closed-loop NetOps/SecOps where agents deliver measurable MTTR and SLO gains.
AI Factories are expensive ecosystems. Design, Deployment, and workload scheduling need to be automated to ensure the desired outcome. This session shows how Aviz ONES delivers the blueprint for SONIC networks and end-end orchestration experience through partner integrations.
Secure the rise of AI agents with zero-trust frameworks, guardrails, and full visibility into machine-driven actions.
As enterprises adopt agentic AI, they face a new class of adversary: autonomous, goal-seeking agents that probe controls, data, and policies at machine speed. Building on ideas like “Digital Darwinism” and adversarial agents competing to optimize infrastructure, this session asks: what happens when attackers weaponize agents—and how do
we respond?
Key Questions: – How will adversarial agents change red-teaming, penetration testing, and threat modeling? – What defensive patterns are emerging for agent endpoint protection, data controls, and PII sensitivity mitigation? – Where do AI guardrails, content moderation, and design best practices fit in the security architecture versus at the application layer? – How do we monitor for agent-vs-agent “arms races” inside the enterprise and prevent unintended escalation?
Takeaways: – A taxonomy of adversarial agent threats relevant to Global 2000 environments. – Concrete examples of guardrail policies and monitoring approaches that actually reduce risk.
This session explores the growing risk posed by quantum-enabled decryption, referred to as Q-Day, the point at which sufficiently powerful quantum computers could undermine widely used public-key cryptosystems. Attendees will examine what this shift means for today’s networks, including the “harvest now, decrypt later” risks and the operational impact of large-scale cryptographic compromise. The discussion will also look at quantum technologies as part of a forward-looking security strategy. Panelists will discuss where these approaches are most relevant, how they can complement broader migration efforts, and what organizations should be doing now to build resilient, tamper-evident communications that can withstand future quantum adversaries.
Who Should Attend:
Cybersecurity professionals, network engineers, IT security architects, and executives responsible for data protection in AI-driven environments. This session is especially valuable for those in finance, healthcare, government, and critical infrastructure sectors where encryption compromise would have severe consequences.
What You Will Learn:
Enterprises are deploying agen
In this session, Rein Security
Using the ONU
Agentic AI systems are useless without provable identity, least-privilege authorization, and strong supervision for non-human actors. This session builds directly on board use cases calling for Zero-Trust identity frameworks for agentic AI in Tier-1/Tier-2 operations, extending those concepts into a full enterprise security model.
Key Questions: – How do you assign and manage identity, credentials, and posture for agents that can provision, patch, and reconfigure networks, applications, and data pipelines? – What does capability-scoped, time-boxed authorization look like in practice for AI tools and agents? – How do we create signed, auditable action trails that satisfy internal audit and regulators for “who/what acted, when, and with what policy”? – How do you integrate agent identity into existing Zero-Trust and privileged access strategies?
Takeaways: – A practical pattern for a Zero-Trust identity framework for agentic AI across network, cloud, and ITSM stacks. – Controls and dashboards that make agent actions observable, provable, and reversible.
This session presents a collaborative approach to building hybrid AI infrastructure that balances public AI services with private model deployment. As AI technology, economics, and governance rapidly evolve in unpredictable ways, the panel will discuss the need for an Enterprise AI Fabric that delivers long-term flexibility and choice. Moderated by Nick Lippis, the session will explore how aligned architectures, data strategies, and operational models can create secure, high-performance environments adaptable to changing AI consumption patterns. Attendees will gain insight into how coordinated multi-vendor ecosystems can future-proof enterprise AI infrastructure.
What You Will Learn
Who Should Attend
This session examines quantum-safe networking as an emerging approach to strengthening security for high-bandwidth connections across large-scale enterprise data center environments. We will define key concepts in quantum-safe networking – including the role of quantum phenomena such as entanglement in supporting secure, tamper-evident communications over fiber infrastructure and hybrid solutions.
With cyber risk increasing and AI workloads driving demand for large-scale data movement, the discussion will explore where conventional cryptographic approaches face operational and lifecycle challenges on high-speed links, and how quantum-safe architectures may complement existing security models. The session will also address practical adoption considerations, including readiness assessments, implementation planning, and hybrid quantum-classical deployment strategies beyond PQC transitions already underway.
Attendees will leave with a clearer understanding of the technology landscape, relevant use cases, and a pragmatic framework for evaluating quantum networking within enterprise infrastructure roadmaps.
Transform operations from NOC to AOC with trusted automation, human-in-the-loop design, and agent-driven workflows at scale.
Enterprises are investing heavily in AI, yet many initiatives stall when moving from pilot to production—not because of models or data, but the network. The connectivity layer remains fragmented, manual, and static, lacking a unified system of truth across locations, providers, and services. This session introduces a programmable connectivity layer for AI-ready infrastructure, combining location intelligence, automated discovery, quote-to-order automation, and continuous reconciliation. By transforming connectivity into a dynamic, software-driven foundation, enterprises can accelerate deployment, improve cost and performance decisions, and enable AI systems to autonomously manage and optimize infrastructure across multi-cloud and edge environments.
Building on the NYC “From NOC to AOC” concept, this session focuses on operational trust: how to evolve from human-centric operations centers to Agentic Operations Centers (AOCs) where AI agents handle most monitoring and remediation—but humans still own accountability. It ties directly into ONUG’s 2026 focus on agentic operations you can trust, including standard identity, guardrails, and observability.
Key Questions: – What work should be fully handed to agents in 2026 (Tier-1, Tier-2 incidents, routine changes), and what must remain human-centric? – How do you train and certify “agent supervisors” to oversee closed-loop systems without being buried in noise? – How do you design human-in-the-loop patterns that are fast enough for AI-era incidents yet still satisfy governance? – What changes in org design, skills, and incentives are needed for teams to trust automation?
Takeaways: – A target operating model for AOCs, including roles, skills, and escalation patterns. – Examples of closed-loop NetOps/SecOps where agents deliver measurable MTTR and SLO gains.
The phrase “network autonomy” is everywhere. But what does it actually look like in practice? In this session, we will demonstrate an interlocking AI architecture working across a real-world network to identify, triage and resolve traditionally hidden issues in minutes, not hours. Watch foundation models detect the anomaly, the reasoning engine correlate it across topology and config state via MCP, and agentic processes propose and execute governed remediation through Ansible. As important, we will introduce a working model for human + AI partnership where every AI decision is transparent, confidence-building and overridable.
This presentation explores the collapse of the legacy, reactive NOC and the rise of the AI-driven Agentic Operations Center (AOC). It outlines how enterprises must strategically delegate operational toil to AI—automating telemetry correlation, triage, and analysis—while preserving human oversight for high-stakes decisions. The session highlights Human-in-the-Loop design, agent-to-agent collaboration, and Zero-Trust guardrails as foundations for safe closed-loop automation. Ultimately, it frames the AOC as both a technological and organizational transformation, elevating engineers from reactive troubleshooting to strategic architecture and autonomous network design.
In multitenant AI clusters, bandwidth is not the SLA—Job Completion Time (JCT) is. In this joint session, NextHop.ai and Keysight present a real-world study of how load balancing modes, ECN thresholds, and congestion control tuning directly impact JCT in shared AI Ethernet fabrics. Using production-class switching and traffic emulation, we demonstrate how overly conservative settings can double completion times and waste GPU cycles—and how systematic tuning makes JCT predictable. Attendees will gain a practical framework for designing and validating AI networks around measurable performance outcomes, not just throughput.
GenAI is powerful in network operations—but it’s probabilistic by nature, meaning it produces “most likely” answers that can vary with small prompt changes, making it risky as a direct control mechanism for repeatable, auditable change workflows. In contrast, deterministic automation (RPA in ops contexts) is built for consistent execution: enforcing config standards, running verified OS upgrade workflows, and remediating drift with predictable outcomes.
In this session, you’ll get a simple decision framework for when to use GenAI vs. deterministic automation, why trusted data (a network source of truth) is a prerequisite for either approach to work reliably at scale, and how to avoid “guardrail sprawl” when teams try to force GenAI into deterministic behavior. We’ll close with the practical operating model that balances speed with control: GenAI accelerates triage and recommendations; humans approve; deterministic automation executes changes with pre-checks and rollback plans.
As AI agents become more capable, a new operational model is emerging: role-based agents designed to support specific operational responsibilities. IT and Network Operations teams interact with agents tailored for frontline responders, specialists, and reliability engineers. These agents combine predictive machine learning, contextual memory, and generative reasoning to summarize incidents, identify root cause, recommend validated remediation steps, and execute approved actions within policy guardrails. In this session, we explore the architecture behind role-based operational agents and how they integrate predictive intelligence, episodic memory, and cross-domain telemetry learning. The discussion previews Grokstream’s L1 Agent for frontline operations and how role-based agents may redefine incident management at scale.
The Special Programs Track highlights the breakthrough technologies and initiatives shaping the future of enterprise IT including SONiC and Quantum Computing
This session explores the growing risk posed by quantum-enabled decryption, referred to as Q-Day, the point at which sufficiently powerful quantum computers could undermine widely used public-key cryptosystems. Attendees will examine what this shift means for today’s networks, including the “harvest now, decrypt later” risks and the operational impact of large-scale cryptographic compromise. The discussion will also look at quantum technologies as part of a forward-looking security strategy. Panelists will discuss where these approaches are most relevant, how they can complement broader migration efforts, and what organizations should be doing now to build resilient, tamper-evident communications that can withstand future quantum adversaries.
Who Should Attend:
Cybersecurity professionals, network engineers, IT security architects, and executives responsible for data protection in AI-driven environments. This session is especially valuable for those in finance, healthcare, government, and critical infrastructure sectors where encryption compromise would have severe consequences.
What You Will Learn:
This session examines quantum-safe networking as an emerging approach to strengthening security for high-bandwidth connections across large-scale enterprise data center environments. We will define key concepts in quantum-safe networking – including the role of quantum phenomena such as entanglement in supporting secure, tamper-evident communications over fiber infrastructure and hybrid solutions.
With cyber risk increasing and AI workloads driving demand for large-scale data movement, the discussion will explore where conventional cryptographic approaches face operational and lifecycle challenges on high-speed links, and how quantum-safe architectures may complement existing security models. The session will also address practical adoption considerations, including readiness assessments, implementation planning, and hybrid quantum-classical deployment strategies beyond PQC transitions already underway.
Attendees will leave with a clearer understanding of the technology landscape, relevant use cases, and a pragmatic framework for evaluating quantum networking within enterprise infrastructure roadmaps.
AI Factories are expensive ecosystems. Design, Deployment, and workload scheduling need to be automated to ensure the desired outcome. This session shows how Aviz ONES delivers the blueprint for SONIC networks and end-end orchestration experience through partner integrations.
eBay operates one of the most advanced enterprise networks built on SONiC and large-scale automation. In this technical deep dive, Rick Casarez provides an inside look at the architecture, design principles, and operational decisions behind eBay’s production SONiC deployment.
Rick will also demonstrate how eBay is applying AI-powered LLMs to assist and automate NOC workflows, enabling engineers to investigate incidents faster, interpret telemetry, and streamline troubleshooting runbooks. The session concludes with a look at how eBay’s network architecture is evolving to support AI workloads and more autonomous network operations.
Key Topics Covered
• Architecture of eBay’s production SONiC-based network
• Design principles behind eBay’s disaggregated networking strategy
• Hardware abstraction and multi-vendor switching architecture
• Automation frameworks and infrastructure-as-code for network operations
• Telemetry, observability, and operational tooling at scale
• Applying LLMs to assist and automate NOC workflows
• AI-assisted troubleshooting, incident investigation, and operational runbooks
• Preparing enterprise networks for AI infrastructure and high-performance workloads
Who Should Attend
• Enterprise network engineers and architects evaluating or deploying SONiC
• NOC and operations leaders exploring AI-assisted operations
• Infrastructure engineers responsible for automation and network reliability
• Platform and SRE teams operating large-scale distributed infrastructure
• IT leaders preparing networks to support AI and high-performance compute environments