New algorithm to improve video game quality

February 18, 2010

Research presented in a paper by Morgan McGuire, assistant professor of computer science at Williams College, and co-author Dr. David Luebke of NVIDIA, introduces a new algorithm to improve computer graphics for video games.

McGuire and Luebke have developed a new method for computerizing lighting and light sources that will allow to approach film quality.

Their paper "Hardware-Accelerated Global Illumination by Image Space Photon Mapping" won a Best Paper award at the 2009 Conference on High Performance Graphics.

Because video games must compute images more quickly than movies, developers have struggled with maximizing graphic quality.

Producing light effects involves essentially pushing light into the 3D world and pulling it back to the pixels of the final image. The method created by McGuire and Luebke reverses the process so that light is pulled onto the world and pushed into the image, which is a faster process.

As video games continue to increase the degree of interactivity, graphics processors are expected to become 500 times faster than they are now. McGuire and Luebke's algorithm is well suited to the quickened processing speed, and is expected to be featured in video games within the next two years.

McGuire is author of "Creating Games: Mechanics, Content, and Technology" and is co-chair of the ACM SIGGRAPH Symposium on Non-Photorealistic Animation and Rendering, and previously chaired the ACM Symposium on Interactive 3D Graphics and Games.

He has worked on and consulted for commercial video games such as "Marvel Ultimate Alliance" (2009), "Titan Quest" (2006), and "ROBLOX" (2005).

Explore further: Samsung Develops Industry’s First Ultra-Fast GDDR4 Graphics DRAM

Related Stories

Samsung Speeds Up World's Fastest Graphics Memory

February 23, 2007

Samsung Electronics announced today that it has increased the data transfer speed of the world’s fastest graphics memory -- GDDR4 (series four of graphics double-data-rate memory) -- by two-thirds. Graphics memory processes ...

Video games shown to improve vision

March 15, 2007

According to a new study from the University of Rochester, playing action video games sharpens vision. In tests of visual acuity that assess the ability to see objects accurately in a cluttered space, game players scored ...

AMD Launches 780 Chipset

March 4, 2008

AMD is announcing the availability of the AMD 780G chipset, designed to deliver the ultimate mainstream computing experience.

Recommended for you

Math reveals unseen worlds of Star Wars

February 10, 2016

Using a new computer program, EPFL researchers offer unusual insight into the universe of Star Wars, which includes more than 20,000 characters spread among 640 communities over a period of 36,000 years.

Tiny diatoms boast enormous strength

February 8, 2016

Diatoms are single-celled algae organisms, around 30 to 100 millionths of a meter in diameter, that are ubiquitous throughout the oceans. These creatures are encased within a hard shell shaped like a wide, flattened cylinder—like ...

Riddle of cement's structure is finally solved

February 9, 2016

Concrete is the world's most widely used construction material, so abundant that its production is one of the leading sources of greenhouse gas emissions. Yet answers to some fundamental questions about the microscopic structure ...

4 comments

Adjust slider to filter visible comments by rank

Display comments: newest first

Scryer
not rated yet Feb 19, 2010
In 10 years, even this feat will look like a small stepping stone... Too bad this technology can't be out sooner.
cybrbeast
not rated yet Feb 19, 2010
"As video games continue to increase the degree of interactivity, graphics processors are expected to become 500 times faster than they are now."

What a useless sentence. It doesn't give a timespan. I could also say GPUs will get a 1000 times faster than they are now. It will happen, the question is when.
degojoey
not rated yet Feb 19, 2010
well 500x in around 15 years, 1000x in 17 years, according to Moore. :D
NanoStuff
not rated yet Feb 21, 2010
1000x in around 10 years, 128,000x in 17 years, according to historical observation.

I also want to emphasize the stupidity of this article, as cybrbeast describes. Also, it's not a new algorithm, it is an existing algorithm running on a GPU.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.