News | Ultrasound Imaging | June 12, 2018 | Tony Kontzer

How AI and Deep Learning Will Enable Cancer Diagnosis Via Ultrasound

NVIDIA GPU Cloud containers speed research into using affordable ultrasound technology more widely

How AI and Deep Learning Will Enable Cancer Diagnosis Via Ultrasound

The red outline shows the manually segmented boundary of a carcinoma, while the deep learning-predicted boundaries are shown in blue, green and cyan. Copyright 2018 Kumar et al. under Creative Commons Attribution License.

June 12, 2018 — Viksit Kumar didn’t know his mother had ovarian cancer until it had reached its third stage, too late for chemotherapy to be effective. She died in a hospital in Mumbai, India, in 2006, but might have lived years longer if her cancer had been detected earlier. This knowledge ate at the mechanical engineering student, spurring him to choose a different path.

“That was one of the driving factors for me to move into the medical field,” said Kumar, now a senior research fellow at the Mayo Clinic, in Rochester, Minn. He hopes that the work his mom’s death inspired will help others to avoid her fate.

For the past few years, Kumar has been leading an effort to use GPU-powered deep learning to more accurately diagnose cancers sooner using ultrasound images. The work has focused on breast cancer (which is much more prevalent than ovarian cancer and attracts more funding), with the primary aim of enabling earlier diagnoses in developing countries, where mammograms are rare.

Into the Deep End of Deep Learning

Kumar came to this work soon after joining the Mayo Clinic. At the time, he was working with ultrasound images for diagnosing pre-term birth complications. When he noticed that ultrasounds were picking up different objects, he figured that they might be useful for classifying breast cancer images.

As he looked closer at the issue, he deduced that deep learning would be a good match. However, at the time, Kumar knew very little about deep learning. So he dove in, spending more than six months teaching himself everything he could about building and working with deep learning models.

“There was a drive behind that learning: This was a tool that could really help,” he said.

And help is needed. Breast cancer is one of the most common cancers, and one of the easiest to detect. However, in developing countries, mammogram machines are hard to find outside of large cities, primarily due to cost. As a result, healthcare providers often take a conservative approach and perform unnecessary biopsies.

Ultrasound offers a much more affordable option for far-flung facilities, which could lead to more women being referred for mammograms in large cities.

Even in developed countries, where most women have regular mammograms after the age of 40, Kumar said ultrasound could prove critical for diagnosing women who are pregnant or are planning to get pregnant, and who can’t be exposed to a mammogram’s X-rays.

Getting Better All the Time

Kumar is amazed at how far the deep learning tools have already progressed. It used to take two or three days for him to configure a system for deep learning, and now takes as little as a couple of hours.

Kumar’s team does its local processing using the TensorFlow deep learning framework container from NVIDIA GPU Cloud (NGC) on NVIDIA TITAN and GeForce GPUs. For the heaviest lifting, the work shifts to NVIDIA Tesla V100 GPUs on Amazon Web Services, using the same container from NGC.

The NGC containers are optimized to deliver maximum performance on NVIDIA Volta and Pascal architecture GPUs on-premises and in the cloud, and include everything needed to run GPU-accelerated software. And using the same container for both environments allows them to run jobs everywhere they have compute resources.

“Once we have the architecture developed and we want to iterate on the process, then we go to AWS [Amazon Web Services],” said Kumar, estimating that doing so is at least eight times faster than processing larger jobs locally, thanks to the greater number of more advanced GPUs in play.

The team currently does both training and inference on the same GPUs. Kumar said he wants to do inference on an ultrasound machine in live mode.

More Progress Coming

Kumar hopes to start applying the technique on live patient trials within the next year.

Eventually, he hopes his team’s work enables ultrasound images to be used in early detection of other cancers, such as thyroid and, naturally, ovarian cancer.

Kumar urges patience when it comes to applying AI and deep learning in the medical field. “It needs to be a mature technology before it can be accepted as a clinical standard by radiologists and sonographers,” he said.

Read Kumar’s paper, “Automated and real-time segmentation of suspicious breast masses using convolutional neural network.”

For more information: www.nvidia.com

This piece originally appeared as a blog post on NVIDIA’s website.

Related Content

Gadolinium contrast agents (GBCAs) are partly retained in the brain, raising safety concerns, as seen in this MRI.

Gadolinium contrast agents (GBCAs) are partly retained in the brain, raising safety concerns, as seen in this MRI.

News | Radiology Business | February 22, 2019
Imaging Technology News has been recognized with three award nominations from the Jesse H.
Welch Road Imaging Integrates RamSoft PowerServer RIS/PACS With openDoctor
News | PACS Accessories | February 20, 2019
Welch Road Imaging in California recently became the first RamSoft customer to integrate openDoctor with its...
Joseph J. Cappello Named Executive Director of Are You Dense Nonprofits
News | Breast Density | February 20, 2019
The boards of Are You Dense Inc. and Are You Dense Advocacy Inc., founded by the late Nancy M. Cappello, Ph.D.,...
Sponsored Content | Videos | Enterprise Imaging | February 20, 2019
At RSNA 2018, Philips Healthcare introduced Performance Bridge as an integral part of its IntelliSpace Enterprise Edi
Congress Directs FDA to Establish Federal Breast Density Inform Standard
News | Breast Density | February 19, 2019 | Jeff Zagoudis, Associate Editor
As part of a federal spending bill passed late Friday, Congress directed the U.S. Food and Drug Administration (FDA) to...
Amazon Comprehend Medical Brings Medical Language Processing to Healthcare
News | Artificial Intelligence | February 15, 2019
Amazon recently announced Amazon Comprehend Medical, a new HIPAA-eligible machine learning service that allows...
Videos | Radiation Therapy | February 15, 2019
ITN Associate Editor Jeff Zagoudis speaks with Vinai Gondi, M.D., director of research and CNS neuro-oncology at the
Fujifilm Exhibits Enterprise Imaging Solutions and Artificial Intelligence Initiative at HIMSS 2019
News | Enterprise Imaging | February 15, 2019
Fujifilm Medical Systems U.S.A. Inc. and Fujifilm SonoSite Inc. showcased their enterprise imaging and informatics...
IBM Watson Health Announces New AI Collaborations With Leading Medical Centers
News | Artificial Intelligence | February 14, 2019
IBM Watson Health announced plans to make a 10-year, $50 million investment in research collaborations with two...
Medivis Launches SurgicalAR Augmented Reality Platform
Technology | Advanced Visualization | February 14, 2019
Medical imaging and visualization company Medivis officially unveiled SurgicalAR, its augmented reality (AR) technology...