Preparing the Network for the Surge in AI Workloads

Now is the time to engage in network planning exercises to address the issues AI workloads and new traffic patterns will have on existing network infrastructures.

Lori MacVittie

September 27, 2024

3 Min Read
Now is the time for network planning to address the issues AI workloads and new traffic patterns will have on existing network infrastructures.
(Credit: genkur / Alamy Stock Photo)

Network newbies, aficionados, and professionals! Lend me your ear. There are things afoot that will have a profound impact on your stomping ground, the network. There is a ton of focus on generative AI and the applications being built to leverage it. Studies abound about what those applications look like, what frameworks they're using, and what services to use. I have some of that data myself.

But what isn’t being talked about (enough) is the impact these applications are going to have on the backbone of business. You know, the network. It wasn’t surprising to learn that only 50% of respondents to a network-oriented survey said network planning was something their company engaged in.

That’s problematic because there are things afoot in business that indicate data center networks—the ones you operate and oversee daily—are about to get inundated with new workloads that will have great implications on data center networks.

AI Workloads Coming Home

Yes, that's what all that talk about repatriation really means. If workloads are coming home, they need a place to stay and hallways to roam around in the middle of the night. Our own research indicates a significant rise (from 13% in 2021 to 50% in 2024) of organizations that have or plan to repatriate workloads.   

And here's the thing: If Barclay’s CIO Survey from the first half of 2024 is correct, the workloads coming home are all about data. Both storage and database workloads were the top workloads moving from the public cloud to a data center near you. Because of the reliance of AI on that data, the return of data-related workloads to the data center is indicative of future AI workloads—in the data center.

Given the current leading AI application patterns, significant changes to the network in terms of throughput and traffic patterns are needed. This is due to the addition of new 'tiers' in the application architecture that expand to incorporate both data sources and inferencing services and an influx of new traffic.

High-Level View of AI-Supporting Architecture

Which Networking Tech is Best for AI Workloads?

The thing is that some protocols, like Infiniband, have long been the purview of storage professionals. But Infiniband is one of the darling protocols of AI factories because most AI traffic is big chunks of data and, honestly, easily overwhelms Ethernet.

So we’ve got a network that’s now incorporating additional protocols with new paths that reach deeper into the data center. There's more traffic and a greater need to control it for various reasons (like compliance and privacy and stuff).

That whole network planning exercise? Absolutely necessary now. And one of the tasks associated with that exercise should be the identification of strategic control points in that network.

Strategic control points are those locations within the network architecture where services can be inserted to control traffic for purposes of scale and security. And just as important as identifying those control points is recognizing that traffic flows are going to be bi-directional. Plenty of AI applications rely on inferencing services that can call out to a service; services that may reside inside the data center but can also be outside the data center.

That’s a marked shift in architecture that has traditionally assumed inbound (ingress) traffic control is the norm. Some of those control points will require a service capable of managing and securing outbound traffic. And not the ‘URL filtering/web content’ kind of capabilities. This is more along the lines of API security turned inside-out.

A Final Word on Networking and AI Workloads

AI is disruptive in every sense of the word. It’s changing industries, it’s changing jobs, it’s changing expectations, and it’s changing the network. When we talk about digital transformation and the need to modernize enterprise architecture, we call out the importance of its most foundational component: infrastructure. That means the network.

It’s time to engage in some network planning exercises before it's overwhelmed with workloads and new traffic requirements that existing network infrastructure can't handle.

Read more about:

Infrastructure for AI

About the Author

Lori MacVittie

Distinguished Engineer, Office of the CTO, at F5

Lori MacVittie is the  Distinguished Engineer in F5's Office of the CTO for cloud computing, cloud and application security, and application delivery and is responsible for education and evangelism across F5's entire product suite. MacVittie has extensive development and technical architecture experience in both high-tech and enterprise organizations. Prior to joining F5, MacVittie was an award-winning Senior Technology Editor at Network Computing Magazine, where she authored articles on a variety of topics aimed at IT professionals. She holds a B.S. in Information and Computing Science from the University of Wisconsin at Green Bay, and an M.S. in Computer Science from Nova Southeastern University. She also serves on the Board of Regents for the DevOps Institute and CloudNOW, and has been named one of the top influential women in DevOps.

SUBSCRIBE TO OUR NEWSLETTER
Stay informed! Sign up to get expert advice and insight delivered direct to your inbox

You May Also Like


More Insights