Computed tomography (CT) is one of the most powerful and well-established diagnostic tools available to modern medicine.
An increasing number of people have been opting for CT scans, raising concerns about the amount of X-ray radiation that patients are exposed to.
Ideally, a patient is exposed to minimum radiation levels during treatments or diagnostic procedures, while still receiving the expected benefit.
In practice, this is known as the ALARA principle, which stands for “As Low As Reasonably Achievable.”
However, this principle results in a trade-off because CT image quality decreases with a decrease in radiation power.
Thus, medical staff usually aim to strike a balance between a patient’s exposure to X-rays and obtaining good quality CT images to avoid misdiagnosis.
This balance can be achieved through an optimisation strategy, in which healthcare professionals, primarily radiologists, observe real images generated by the tomographer and try to identify features, such as tumours or abnormal tissue.
Following this, a specialist employs statistical methods to calculate the optimal radiation dose and configuration of the tomographer.
This procedure can be generalised by employing reference CT images obtained by scanning specifically designed phantoms containing inserts of different sizes and contrasts, which represent standardised abnormalities.
Nevertheless, such manual image analyses are very time-consuming.
To address this issue, a team of researchers from Italy led by Dr Sandra Doria and members of the Physics Department at the University of Florence, in collaboration with radiologists and medical physicists from Florence Hospital, explored the possibility of automating this process using artificial intelligence (AI).
As reported in Journal of Medical Imaging, the team created and trained an algorithm—a “model observer”—based on convolutional neural networks (CNNs), which could analyse the standardised abnormalities in CT images just as well as a professional.
To do so, the team had to generate enough training and testing data for the model.
Thirty healthcare professionals visually examined 1000 CT images, each consisting in a phantom that mimics human tissue.
Aptly termed “phantom,” this material contained cylindrical inserts of different diameters and contrasts.
The observers were asked to identify if and where the inserted object appeared in each of the images and state how confident they were in their assessment.
This resulted in a dataset of 30,000 labelled CT images taken using different tomographic reconstruction configurations, accurately reflecting human interpretation.
Next, the team implemented two AI models based on different architectures—UNet and MobileNetV2.
They modified the base design of these architectures to enable them to perform both classification (“Is there an unusual object in the CT image?”) and localisation (“Where is the unusual object?”).
Then, they trained and tested the models using images from the dataset.
Through statistical analyses, the research team evaluated various performance metrics to verify that the model observers could accurately emulate how a human would assess the CT images of the phantom.
“Our results were very promising, as both trained models performed remarkably well and achieved an absolute percentage error of less than 5 percent. This indicated that the models could identify the object inserted in the phantom with similar accuracy and confidence as a human professional, for almost all reconstruction configurations and abnormalities sizes and contrasts,” remarked Doria, while discussing their findings.
Doria and her team believe that with additional efforts, their model could become a viable strategy to automatically assess CT image quality.
She further adds, “Our CNN-based model observers could greatly simplify the process of optimising the radiation dose used in CT protocols, thereby minimising health risks to the patient, and help avoid the time-consuming limitations of medical evaluations.”
Doria expressed confidence that the team will succeed in applying their AI model observers on a larger scale, making CT evaluations faster and safer than ever before.
We are an independent charity and are not backed by a large company or society. We raise every penny ourselves to improve the standards of cancer care through education. You can help us continue our work to address inequalities in cancer care by making a donation.
Any donation, however small, contributes directly towards the costs of creating and sharing free oncology education.
Together we can get better outcomes for patients by tackling global inequalities in access to the results of cancer research.
Thank you for your support.