AI Applications Put More Focus on the Network
AI is going to impact the network in more ways than one. The expansion of the N-S traffic path and the fight for AI compute interconnection are two of the most important right now.
June 19, 2024
There’s a lot going on with AI right now. You can’t read an article or blog without it being mentioned. A lot of the chatter around AI is focused on the user experience and how it will change our lives and jobs from a productivity standpoint. This is all true, but there isn't as much chatter about the impact AI applications are having on the network. And trust me, it is.
Our most recent research indicates organizations are allocating, on average, 18% of their IT budget just for AI. Of that, 18% is going to models, and 9% is being spent on GPUs. Organizations are clearly investing in building out their AI capabilities, with a healthy percentage (60%) buying up GPUs to put on-premises. That implies some rather significant changes to the data center.
That’s probably not a surprise because the entire data center architecture is being impacted by AI. Specifically, by the way AI compute complexes are being designed and built out (that’s what those GPUs are for, after all). There’s a lot going on, including a new spat brewing between, well, everyone and NVIDIA. It’s akin to the protocol wars that led to TCP/IP becoming the standard for networking in the data center.
And while this spat is focused on the GPU interconnects within an AI “pod” (basically a Kubernetes cluster comprised of inferencing servers and a lot of GPUs and CPUs and other PUs) there’s also the new interconnect between these AI compute complexes and the rest of the data center.
AI drives data center bifurcation
The data center is bifurcating into two distinct computing complexes; one focused on applications (that may or may not use AI) and an AI compute complex in which inferencing is run at scale.
There’s a lot to unpack here, not the least of which is the introduction of a second N-S insertion point inside the data center for AI traffic. That’s the point at which traffic moves from the existing data center to the AI compute complex. Like any new interconnect, there are a number of network and application layer needs that must be addressed, such as:
Dynamic routing integration for resiliency and service load distribution
Per-tenant network isolation for ingress and egress for any K8s workload
Advanced CGNAT capabilities
Per-tenant network firewalling
SIEM integration
IDS services
L4-L7 SLA and throttling
There are all the expected network services required for traffic traversing two 'environments' as well as the helpful higher-layer network (L4-7) traffic management to keep inferencing servers from being overwhelmed by requests.
Notice most of these aren’t your typical “ingress control” functions, despite the ten-thousand-foot view that traffic is likely passing from one Kubernetes cluster to another. The interconnect here requires something more robust than just routing. The focus is clearly on network capabilities common to most data center interconnects.
Per-tenant network isolation may not be a typical enterprise requirement, but AI is making it more than just a service and cloud provider necessity. Per-tenant based networking and architecture constructs will be increasingly important in the AI era to enable enterprises to prioritize AI workloads—like automation and operational analytics—to make sure they are not suffocated by lower-priority AI workloads.
I’ve talked (and talked and talked) about the need to modernize infrastructure in the past. That includes—and indeed focuses on—the network. There are no applications, no tools, or technologies that do not rely on the network today. All applications, and especially AI applications, need a flexible foundation on which to deliver their digital value to users. That foundation is the network, and it's one of the reasons we tag infrastructure distributedness and automation as key technical capabilities needed for any organization to become a digital business.
AI is going to impact the network in more ways than one. But the expansion of the N-S traffic path and the fight for AI compute interconnection are two of the most important right now. Because together they will determine the direction of next generation data center architectures.
Related articles:
Read more about:
Infrastructure for AIAbout the Author
You May Also Like