Using technology similar to the type that powers facial and speech recognition on a smartphone, researchers at the University of North Carolina Lineberger Comprehensive Cancer Centre have trained a computer to analyse breast cancer images and then classify the tumours with high accuracy.
In a study published in the journal NPJ Breast Cancer, researchers reported they used a form of artificial intelligence called machine learning, or deep learning, to train a computer to identify certain features of breast cancer tumours from images.
The computer also identified the tumour type based on complex molecular and genomic features, which a pathologist can't yet identify from a picture alone.
They believe this approach, while still in its early stages, could eventually lead to cost savings for the clinic and in breast cancer research.
"Your smartphone can interpret your speech, and find and identify faces in a photo," said the study's first author Heather D. Couture, a graduate research assistant in the UNC-Chapel Hill Department of Computer Science. "We're using similar technology where we capture abstract properties in images, but we're applying it to a totally different problem."
For the study, the researchers used a set of 571 images of breast cancer tumours from the Carolina Breast Cancer Study to train the computer to classify tumours for grade, oestrogen receptor status, PAM50 intrinsic subtype, histologic subtype, and risk of recurrence score.
To do this, they created software that learned how to predict labels from images using a training set, so that new images could be processed in the same way.
They then used a different set of 288 images to test the computer's ability to distinguish features of the tumour on its own, comparing the computer's responses to findings of a pathologist for each tumour's grade and subtype, and to separate tests for gene expression subtypes.
They found the computer was able to distinguish low-intermediate versus high-grade tumours 82 percent of the time.
When they had two pathologists review the tumour grade for the low-intermediate grade group, the pathologists agreed with each other about 89 percent of the time, which was slightly higher than the computer's accuracy.
In addition, the computer identified oestrogen receptor status, distinguished between ductal and lobular tumours, and determined whether each case had a high or low risk of recurrence high levels of accuracy.
It also identified one of the molecular subtypes of breast cancers - the basal-like subtype - which is based on how genes within the tumour were expressed - with 77 percent accuracy.
"Using artificial intelligence, or machine learning, we were able to do a number of things that pathologists can do at a similar accuracy, but we were also able to do a thing or two that pathologists are not able to do today," said UNC Lineberger's Charles M. Perou, PhD, the May Goldman Shaw Distinguished Professor of Molecular Oncology, professor of genetics, and of pathology and laboratory medicine in the UNC School of Medicine. "This has a long way to go in terms of validation, but I think the accuracy is only going to get better as we acquire more images to train the computer with."
The computer's ability to identify the basal-like subtype was exciting to researchers, and could have applications in cancer research.
They also believe the technology could have applications in communities that do not have pathology resources as well as in helping to validate pathologists' findings.
"We were surprised that the computer was able to get a pretty high accuracy in estimating biomarker risk just from looking at the pictures," said UNC Lineberger's Melissa Troester, PhD, a professor in the UNC Gillings School of Global Public Health. "We spend thousands of dollars measuring those biomarkers using molecular tools, and this new method can take the image and get 80 percent accuracy or better at estimating the tumour phenotype or subtype. That was pretty amazing."
Couture said deep learning technology has been used in a range of applications, including speech recognition and autonomous vehicles.
"Humans can look at one or two examples of something, and be able to generalise when they see other objects," Couture said. "For example, chairs come in so many different forms, but we can recognise it as something we sit on. Computers have a much harder time generalising from small amounts of data. But on other hand, if it you provide enough labeled data, they can learn concepts that are much more complex than humans can assess visually - such as identifying the basal-like subtype from an image alone."
The unique aspect of their work, researchers said, was that they were able to use the technology to see features of the tumours that humans cannot.
They want to figure out what the computer is seeing, as well as to study whether the technology could predict outcomes.
"The computer extracted a lot of information from the images," Troester said. "We would like to test how well these features predict outcomes, and if we can use these features together with things like molecular data to do even better at giving patients a precise view of what their disease course looks like, and what treatments might be effective."
We are an independent charity and are not backed by a large company or society. We raise every penny ourselves to improve the standards of cancer care through education. You can help us continue our work to address inequalities in cancer care by making a donation.
Any donation, however small, contributes directly towards the costs of creating and sharing free oncology education.
Together we can get better outcomes for patients by tackling global inequalities in access to the results of cancer research.
Thank you for your support.