Computing's invisible challenge

April 8, 2014 by Angela Herring
Assistant professor of electrical and computer engineering Ningfan Mi was recently named one of 42 recipients of the Air Force Office of Scientific Research Young Investigator Award. Photo by Brooks Canaday

To us, it may not seem like a big deal: CNN's website is taking too long to load. The day's most popular YouTube video won't stop buffering. "Twitter is over capacity." While these little hiccups in usability may frustrate end users, they merely scratch the surface of the enormous technical challenge that's confronting the backend.

Northeastern University assistant professor of electrical and computer engineering Ningfang Mi recently learned she was one of 42 early-career researchers to win a Young Investigator Award from the Air Force Office of Scientific Research. They will receive the grants over a three-year period.

She plans to use award to figure out a better way to manage the vast amount of information sharing that takes place online—and push that massive even further into the background for end users.

These days most of the data we request online is stored in the so-called "cloud"—a series of virtual computers distributed on physical servers around the world. For instance, Google has 12 data centers across four continents. The 20,000 emails sitting in my Gmail inbox aren't actually stored on my computer—they're stored in Google's cloud, which exists on all those remote servers. Every time I look at one of my emails, I am requesting access to it from one of those servers.

Now consider YouTube. Its billions of hours of video aren't all sitting on the same physical server; rather, they are stored remotely in the cloud. In this case, I am just one of millions of users requesting the same video in a given moment. And that, Mi explained, is where things get challenging.

Her research is focused on modeling performance in different scenarios and figuring out the best ways to manage resources based on the outcomes of those models. This will give her a sense of the workloads and number of traffic requests that remote servers are likely to have to handle.

"Based on this kind of information," she said, "how can I find the best configuration for the platform in order to provide the highest quality of service?"

There are two options: She can either move information around on a single server or move information between servers. The best choice will depend on the situation at hand.

"Before predictions were based more on average load or traffic, but now we know that in reality the workload changes," Mi said. "The term I use here is 'burstiness' or 'spikes.'"

Indeed, it all depends on the burstiness of human behavior. Some online phenomena are predictable, Mi said. For instance, you're likely to see a burst in email activity on the East Coast every weekday at around 9 a.m. EST. Similarly, the Internet is likely to be all-a-flurry across a range of websites on election night as people world over discuss the race on Twitter, stream acceptances speeches on NBC, and read about the results in The New York Times.

But what about when a celebrity unexpectedly passes away or makes a comment that goes viral? Or when a boy in a balloon suddenly becomes one of the biggest news stories on the Internet? No one can predict events like that, so no amount of resource management preparation could ready YouTube for the associated activity spikes.

Mi, for her part, is developing models that will help detect those bursts with more immediacy—and in some cases even predict them a couple hours in advance. So while we may not know when the next media hoax will drive traffic from millions of curious viewers, at least our computers will be able to handle it better.

Explore further: Google sits on Internet-traffic throne, says report

Related Stories

Google sits on Internet-traffic throne, says report

July 23, 2013

The stats are in. According to Deepfield, an Internet analytics business, Google serves 25 percent of North American Internet traffic. The question becomes, glass half empty or full and overflowing?

IBM to invest $1b in Linux, open-source

September 17, 2013

IBM said Tuesday it would invest $1 billion in new Linux and open source technologies for its servers in a bid to boost efficiency for big data and cloud computing.

Making elastic cloud computing a reality

April 8, 2014

( —University of New South Wales researchers are using artificial intelligence to create a computer network capable of regulating its own consumption of public cloud services.

Recommended for you

Using stories to teach human values to artificial agents

February 12, 2016

The rapid pace of artificial intelligence (AI) has raised fears about whether robots could act unethically or soon choose to harm humans. Some are calling for bans on robotics research; others are calling for more research ...

AT&T testing superfast 5G mobile network

February 12, 2016

AT&T announced plans Friday to begin testing a superfast fifth generation, or 5G, mobile network with speeds up to 100 times faster than the most rapid existing connections.

Tiny diatoms boast enormous strength

February 8, 2016

Diatoms are single-celled algae organisms, around 30 to 100 millionths of a meter in diameter, that are ubiquitous throughout the oceans. These creatures are encased within a hard shell shaped like a wide, flattened cylinder—like ...


Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.