Nvidia buys ray-tracing tech company RayScale

Graphics chipmaker acquiring small company that is a product of 10 years of research at the University of Utah.

Nvidia confirmed Friday that it has acquired RayScale, a small company that develops ray-tracing technology. Financial terms of the deal have not been disclosed.

Ray tracing has been mentioned frequently by Intel over the last six months. An Intel blog titled "Real Time Ray-Tracing: The End of Rasterization?" and later comments by Intel executives that the company is looking at doing ray tracing on its processors set the stage for debate on the viability of ray tracing in mainstream gaming.

PC graphics technology today uses rasterization. (A discussion of ray tracing vs. rasterization.)

Ray Tracing is a technique for rendering three-dimensional graphics with extremely complex light interactions, allowing the creation of transparent surfaces and shadows, for example, with stunning photorealistic results.

Ray tracing is a highly parallel process. And the GPU (graphics processing unit) provides high level of parallelism, according to Nvidia officials speaking at a conference on Thursday. The GPU has special function units that were desgined for doing graphics operations that are perfect for ray tracing, said Nvidia Chief Scientist David Kirk.

At the conference, Kirk and RayScale scientists discussed "GPU ray tracing." It's not clear how soon this technology would be used commercially by Nvidia.

RayScale, which provides interactive ray tracing and photo-realistic rendering solutions, says its technologies "dramatically increase the speed and realism at which graphics professionals can produce high quality three-dimensional computer graphics and photorealistic computer images."

RayScale is a product of the decade-long interactive ray-tracing research at the University of Utah, according to RayScale.

At the Intel Developer Forum in Shanghai in April, Senior Intel Vice President Patrick Gelsinger spelled out Intel's vision: essentially that ray tracing-based rendering technologies can be used in high-end gaming. "An intro of these capabilities into mainstream gaming we believe is possible in the future," Gelsinger said at that time.

Intel though has provided little proof beyond the running of some aspects of Quake IV (a relatively old game), for example, on an eight-core system.

Featured Video
This content is rated TV-MA, and is for viewers 18 years or older. Are you of age?
Sorry, you are not old enough to view this content.

Is 'Chipgate' the new iPhone controversy?

We survived "Bendgate" with the iPhone 6 -- is it "Chipgate" for the iPhone 6S? Plus, you can expect the new iPad Pro and Apple TV by early November.

by Brian Tong