Content

Image
Shlomo Zilberstein and Neil Immerman
Shlomo Zilberstein (left) and Neil Immerman (right).

Training one autonomous car to make safe driving decisions is already challenging. Add other self-driving cars to the road—each operating with different information—and the decision-making problem becomes dramatically more complex.  

That is the core finding of a 2002 research paper by Manning College of Information and Computer Sciences (CICS) alumnus Daniel S. Bernstein ’05PhD, Professor Emeritus Neil Immerman, Professor Shlomo Zilberstein, and Robert Givan of Purdue University. The research was recently selected as a landmark paper for the Mathematics of Operations Research journal’s 50th anniversary, which highlighted one paper from each year of its history. 

The paper, “The Complexity of Decentralized Control of Markov Decision Processes,” demonstrated that coordinating decisions among two or more agents with limited information—such as multiple self-driving cars—is substantially more complex than for a single agent. Since its publication, the work has influenced continued research into how multiple agents can coordinate decisions under uncertainty when communication between them is limited. 

“This work showed that the true source of difficulty in multi-agent decision-making is not just uncertainty, but decentralization—multiple agents acting with different, limited views of the world,” Zilberstein said. “By isolating this source of complexity, it gave researchers a deeper understanding of why these problems are so hard, and how to design principled approximation and coordination methods that work well in practice despite the underlying complexity.” 

Markov decision processes, or MDPs, model decision-making for an agent that knows the current state of its environment. For example, an MDP might model how a robot chooses its actions step-by-step based on what it knows about the current state of the world. 

The challenge becomes much greater when multiple agents must make decisions without central coordination. In their 2002 paper, the research team analyzed this framework, known as decentralized control, and mathematically proved that computing optimal decision-making strategies is vastly more complex than in the single-agent case. 

“In systems like autonomous vehicles, each car is an agent with partial information—communication is limited or unreliable, and safety depends on coordinated behavior,” Zilberstein said. “This work told us not to expect perfect coordination in general, but instead to focus on approximation methods that produce robust policies.” 

Today, the team’s results have influenced artificial intelligence research on planning, deep multi-agent reinforcement learning, and robotics.  

“Thanks to this work, researchers have a precise mathematical model for decentralized decision-making,” Zilberstein said. “It became clear that decentralization—even involving just two agents—is not a small complication; it creates a fundamentally different class of problems. Realizing that led to the development of rich new methods to combat this complexity shift, which in some cases resulted in better ways to tackle single-agent decision-making.” 

The anniversary recognition reflects the paper’s foundational and enduring role in defining a new research direction that has influenced artificial intelligence, robotics, and reinforcement learning. In addition to this recognition, the paper received an Influential Paper Award from the International Foundation for Autonomous Agents and Multiagent Systems in 2019

Article posted in Research