Computing's invisible challenge

Apr 08, 2014 by Angela Herring
Assistant professor of electrical and computer engineering Ningfan Mi was recently named one of 42 recipients of the Air Force Office of Scientific Research Young Investigator Award. Photo by Brooks Canaday

To us, it may not seem like a big deal: CNN's website is taking too long to load. The day's most popular YouTube video won't stop buffering. "Twitter is over capacity." While these little hiccups in usability may frustrate end users, they merely scratch the surface of the enormous technical challenge that's confronting the backend.

Northeastern University assistant professor of electrical and computer engineering Ningfang Mi recently learned she was one of 42 early-career researchers to win a Young Investigator Award from the Air Force Office of Scientific Research. They will receive the grants over a three-year period.

She plans to use award to figure out a better way to manage the vast amount of information sharing that takes place online—and push that massive even further into the background for end users.

These days most of the data we request online is stored in the so-called "cloud"—a series of virtual computers distributed on physical servers around the world. For instance, Google has 12 data centers across four continents. The 20,000 emails sitting in my Gmail inbox aren't actually stored on my computer—they're stored in Google's cloud, which exists on all those remote servers. Every time I look at one of my emails, I am requesting access to it from one of those servers.

Now consider YouTube. Its billions of hours of video aren't all sitting on the same physical server; rather, they are stored remotely in the cloud. In this case, I am just one of millions of users requesting the same video in a given moment. And that, Mi explained, is where things get challenging.

Her research is focused on modeling performance in different scenarios and figuring out the best ways to manage resources based on the outcomes of those models. This will give her a sense of the workloads and number of traffic requests that remote servers are likely to have to handle.

"Based on this kind of information," she said, "how can I find the best configuration for the platform in order to provide the highest quality of service?"

There are two options: She can either move information around on a single server or move information between servers. The best choice will depend on the situation at hand.

"Before predictions were based more on average load or traffic, but now we know that in reality the workload changes," Mi said. "The term I use here is 'burstiness' or 'spikes.'"

Indeed, it all depends on the burstiness of human behavior. Some online phenomena are predictable, Mi said. For instance, you're likely to see a burst in email activity on the East Coast every weekday at around 9 a.m. EST. Similarly, the Internet is likely to be all-a-flurry across a range of websites on election night as people world over discuss the race on Twitter, stream acceptances speeches on NBC, and read about the results in The New York Times.

But what about when a celebrity unexpectedly passes away or makes a comment that goes viral? Or when a boy in a balloon suddenly becomes one of the biggest news stories on the Internet? No one can predict events like that, so no amount of resource management preparation could ready YouTube for the associated activity spikes.

Mi, for her part, is developing models that will help detect those bursts with more immediacy—and in some cases even predict them a couple hours in advance. So while we may not know when the next media hoax will drive traffic from millions of curious viewers, at least our computers will be able to handle it better.

Explore further: Making elastic cloud computing a reality

add to favorites email to friend print save as pdf

Related Stories

Making elastic cloud computing a reality

Apr 08, 2014

(Phys.org) —University of New South Wales researchers are using artificial intelligence to create a computer network capable of regulating its own consumption of public cloud services.

Google sits on Internet-traffic throne, says report

Jul 23, 2013

The stats are in. According to Deepfield, an Internet analytics business, Google serves 25 percent of North American Internet traffic. The question becomes, glass half empty or full and overflowing?

IBM to invest $1b in Linux, open-source

Sep 17, 2013

IBM said Tuesday it would invest $1 billion in new Linux and open source technologies for its servers in a bid to boost efficiency for big data and cloud computing.

Recommended for you

Forging a photo is easy, but how do you spot a fake?

Nov 21, 2014

Faking photographs is not a new phenomenon. The Cottingley Fairies seemed convincing to some in 1917, just as the images recently broadcast on Russian television, purporting to be satellite images showin ...

Algorithm, not live committee, performs author ranking

Nov 21, 2014

Thousands of authors' works enter the public domain each year, but only a small number of them end up being widely available. So how to choose the ones taking center-stage? And how well can a machine-learning ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.