.

Arpit Gupta Receives Google Funding to Develop Network Foundation Models

Thursday, July 24, 2025

UC Santa Barbara computer science assistant professor Arpit Gupta has been recognized with two highly competitive research awards from Google, underscoring his pioneering work at the intersection of machine learning and computer networks. Gupta received a Google Research Scholar Award, part of a prestigious program that supports early-career faculty pursuing promising and impactful research. He was also named an inaugural recipient of the Google ML (machine learning) and Systems Junior Faculty Award, which was launched this year to recognize exceptional junior faculty worldwide who are conducting research at the interface of machine learning and systems. The award is considered one of the most selective recognitions for early-career faculty in this field, with the inaugural cohort of recipients drawn from top institutions around the nation. Together, the two awards, totaling $160,000, place Gupta among an elite group of emerging leaders who are shaping the future of intelligent networks.

"The award is going to more than fifty assistant professors in twenty-seven U.S. universities whose research is particularly noteworthy for Google,” said Amin Vahdat, VP/GM of ML, Systems & Cloud AI. “These professors are leading the analysis, design, and implementation of efficient, scalable, secure, and trustworthy computing systems. Their work crosses the technology stack, from algorithms to software and hardware, enabling machine learning and cloud computing at an increasingly massive scale.”
The Research Scholar Award is similarly competitive, supporting groundbreaking research that could have a broad impact across disciplines. For Gupta, the awards validate his bold vision that involves rethinking how networks can be made “self-driving” by using a new class of machine-learning models capable of managing themselves with minimal human intervention.

For decades, machine-learning problems in networking have been solved through point solutions — single models tailored to solve individual problems. But as networks have grown larger and more complex, maintaining separate models for each decision task has become time-consuming, computationally expensive, and difficult to scale.

Gupta is pursuing a new approach he calls the convergence principle, his term for the idea that, instead of building and maintaining separate models for each task, it may be possible to develop a single, general-purpose foundation model — a large, general-purpose model pre-trained on diverse data and fine-tuned for specific tasks as needed — that can adapt to a wide variety of networking problems across different scales.

“Over time, my research has been guided by this convergence principle,” Gupta explains. “We started to ask: if we want to build self-driving networks capable of making a wide range of decisions, do we really need to engineer thousands of distinct machine-learning models, or is there a smarter way to unify these capabilities in a single system?”

His work is inspired by the trajectory of the natural language processing (NLP) community, which faced similar challenges in developing task-specific models. Initially, NLP researchers built separate models for translation, sentiment analysis, and question answering, each trained and maintained independently. The turning point came with the development of foundation models, such as BERT in 2018 and later GPT-3, which demonstrated that training on broad, diverse datasets could yield a single adaptable model. “They showed it was possible to move beyond point solutions by building a foundation model that could be fine-tuned for a variety of tasks,” Gupta notes. “We’re exploring what it would take to bring that approach to networking.”

But networking, Gupta emphasizes, poses unique challenges. “We wondered whether we could borrow what has been done in other areas and adapt it for networking, or if we would need to start from scratch and build something specific to the networking environment [a phrase referring not only to the static infrastructure but also to the dynamic, constantly changing context [which includes traffic patterns, interference, competing user demands, and even malicious activity],” he says. “We realized that, while there are similarities in what makes a good model, networking has its own unique problems.”

One of Gupta’s research directions focuses on analyzing packet traces, which capture how network rules and protocols interact with that dynamic environment. By examining patterns in these packets — the smallest units of data on a network — Gupta can infer information about the state of individual hosts, subnets, and even the broader network. “When I combine groups of packets sent by multiple connections on the same host, it tells me something about the host itself or the group of devices in that subnet,” he explains.

Another challenge unique to networking is scale. Modern networks can transmit billions of packets every second, making it impossible to observe and process each one individually. “That problem of scale has always been there,” Gupta says, “and it means we have to build solutions that are smart enough so that we can avoid processing every packet individually.”

To address this challenge, Gupta’s group is developing what he calls a selective representation approach. Every packet that travels over a network carries two parts: a header and a payload. The header contains routing information, such as the source and destination, while the payload holds the actual data being sent. Deciding how much of this information to examine becomes a balancing act, Gupta explains: “We need to figure out how much information is enough to make an informed decision without incurring enormous computational costs. One way to do that is by designing intelligence that minimizes the amount of data extracted from each packet while maximizing the overall insight we gain.”

In this approach, rather than trying to inspect every packet, the system identifies representative packets to act as proxies for the rest, reducing the computational burden while accessing essential information. “A very important aspect of networking is that we can’t treat all packets the same way,” Gupta says. “What you do depends on the type of information you need. Counting packets is computationally cheap, but understanding where a packet is going and who sent it requires inspecting the headers, which is more expensive. And if you need to examine the payload — the actual content — that’s even more intensive.”

Gupta’s broader vision is to create a foundation model for networks that can leverage multi-modal data from diverse sources — including packet traces, telemetry logs, and device statistics — to solve complex learning problems at different scales. This could mean making decisions every few packets, every second, or every few minutes — a flexibility not feasible with today’s task-specific models.

Currently, many networks rely on stop-gap solutions that are trained and optimized for individual use cases. While effective for narrow tasks, this approach doesn’t scale well. “Developing a task-specific model for each learning problem is not sustainable,” Gupta says. “It takes significant effort to design, train, and optimize each one. We believe a foundation model is the next step.”

Gupta is already working with ESnet, a national research network, to explore how such a model might be deployed. “There’s already momentum,” he says. “Companies like Cisco and others are paying attention to what this solution is going to look like. I think in the next few years, a foundation model for networking is going to be a real thing.”

Image of Arpit Gupta

Arpit Gupta, photograph by Lilli Walker.