Gaming tech spurs development of hyperfast medical imaging systems

University, hospital team to build tools to boost medical image processing by 1,000 times.

University and hospital researchers have taken a cue from the gaming industry by using 3D video graphics chips to develop a parallel software platform that can speed the processing times, in this case for medical digital imaging, by 10 to 1,000 times.

Northeastern University in Boston and Massachusetts General Hospital (MGH) were jointly awarded a $1.3 million grant from the National Science Foundation in December to develop the technology and use it to enhance several biomedical imaging applications, including software designed for breast and brain imaging.

Northeastern, which is working with several other schools, expects to be able to deliver software for medical vendors by year's end.

Early on in the effort, MGH has already shown that the parallel technology can cut the time it takes to read a 3D-radiological breast image from hours to seconds, meaning a patient could be cleared to have a biopsy during a visit to the doctor for a breast exam.

"That way they can determine whether there's a malignant tumor in the breast," said Professor Dave Kaeli, who leads Northeastern's parallel software development team. "The implications for a patient's mental health as well as cost of care for patients are huge,"

The new software libraries will enable radiologists to more quickly assess tissue images, which will allow clinicians to perform image-guided biopsies earlier than can be done now. The faster process will lead to quicker cancer diagnoses detection of cardiovascular plaque. The software will also allow for the rapid delivery of data needed by surgeons; and for the delivery of more accurate radiation treatment.

Kaeli's team is developing the software on systems running Graphics Processing Units (GPUs), very high-performance processors with as many as 200 cores on a single chip.

The development team is using proprietary development software, called CUDA, from GPU maker Nvidia, and the open source OpenCL, Open Computing Language, that was originally developed by Apple Inc. Cuda and OpenCL act as a framework for writing software programs that can execute across heterogeneous GPU platforms.

"We can use one or a few cards. What we do is take a serial application and rewrite it to exploit the data-level parallelism available on graphics cards," Kaeli said.

Homer Pien, PhD, director of the laboratory for medical imaging and computation in MGH's Department of Radiology, said the parallel processing software will enable the use of a radiological imaging technique called iterative reconstruction. The technique can complete x-ray tracing in the same way computer graphics in the gaming industry uses light ray tracing as a technique for generating an image by tracing the path of light.

Iterative reconstruction is capable of producing a very high degree of realism, but is CPU intensive, thus the need for the GPUs and the new parallel processing technology. Iterative reconstruction will afford higher resolution images than today's CT scans, PET scans and MRIs, but with lower dosages of X-rays.

"We could also end up with images that have higher resolution, lower signal to noise ratio and lower artifacts," said Pien, whose team is developing the algorithms necessary for software to behave properly for medical purposes.

An artifact is an object in an x-ray image, such as plaque in an artery, that can be distorted and appear larger than it is.

"We tend to think of the medical enterprise as this big monolithic thing that feeds itself, but it turns out we're actually benefiting from this whole gaming revolution," Pien said.

Currently, three Northeastern professors and 10 graduate students are working on developing the parallel software to work on GPU-powered computer systems. The team is hoping to deliver its first software platform build to develop new medical imaging technology by the end of this year.

The Northeastern University team is also working with MGH's Radiation Physics Department to develop an image-guided radiotherapy system that can process images of moving tumors in real-time to precisely direct radiation theoropy.

"You're trying to [get] the device to only follow the tumor, and that's a very compute-intensive application that's generally done offline," he said. "This improves the outcomes of cancer patients."

The researchers are also working with surgery simulation manufacturer SimQuest Corp. to enhance systems that military doctors use to train for performing surgery to remove shrapnel from soldiers hit by road-side bombs in Iraq and Afghanistan, Kaeli said.

"This whole area is very exciting right now," he said.

Lucas Mearian covers storage, disaster recovery and business continuity, financial services infrastructure and health care IT for Computerworld. Follow Lucas on Twitter at @lucasmearian, send e-mail to lmearian@computerworld.com or subscribe to Lucas's RSS feed .

Join the discussion
Be the first to comment on this article. Our Commenting Policies