Researchers from Intermountain Healthcare and Stanford University say 10 seconds is about how quickly it took a new system they studied that utilizes artificial intelligence to accurately identify key findings in chest x-rays of patients in the emergency department suspected of having pneumonia.
The study found that those ultra-quick findings may enable physicians reading x-rays to accurately confirm a pneumonia diagnosis significantly faster than current clinical practice, enabling treatment to start sooner, which is vital for severely ill patients who’re suffering from pneumonia. Findings from the collaborative study were presented at the European Respiratory Society’s International Congress 2019, held in Spain in late September.
Researchers from Intermountain and Stanford studied the CheXpert system, an automated chest x-ray interpretation model developed at Stanford University that utilizes artificial intelligence, to review x-ray images taken at a number of emergency departments at Intermountain hospitals throughout Utah. Upon review, researchers found the CheXpert system identified key findings in x-rays very accurately—with high agreement to a consensus of three radiologists—in about 10 seconds, which significantly outperforms current clinical practice.
“CheXpert is going to be faster and as accurate as radiologists viewing the studies. It’s an exciting new way of thinking about diagnosing and treating patients to provide the very best care possible,” says Nathan C. Dean, MD, principal investigator of the study, and section chief of pulmonary and critical care medicine at Intermountain Medical Center in Salt Lake City.
The CheXpert model was developed by the Stanford Machine Learning Group, which used 188,000 chest imaging studies to create a model that can determine what is and is not pneumonia on an x-ray. These images were taken from the Stanford Medical Center in Palo Alto, Calif.
Since patient populations are different per geographic locations, CheXpert was then fine-tuned for Utah by reading an additional 6,973 images from Intermountain emergency departments.
“We’ve been developing a deep learning algorithm that can automatically detect pneumonia and related findings in chest x-rays,” says Jeremy Irvin, a PhD student at Stanford, and member of the research team. “In this initial study, we’ve demonstrated the algorithm’s potential by validating it on patients in the emergency departments at Intermountain Healthcare. Our hope is that the algorithm can improve the quality of pneumonia care at Intermountain, from improving diagnostic accuracy to reducing time to diagnosis.”
In a typical emergency department, Dean explains, patients suspected of having pneumonia get a chest x-ray. While creating those images is a quick process, having them read can be time-consuming since those x-rays go into a line with other images to be interpreted by radiologists. That process can take to 20 minutes or more, which means potential delays in the start of antibiotics for very sick pneumonia patients.
At Intermountain emergency departments, radiology reports are run through the Cerner Natural Language Processing (NLP), which is a support tool currently used to get needed information from the radiologist report. NLP then feeds the information into ePNa, an electronic clinical decision support tool part of usual pneumonia care at Intermountain.
For most emergency departments where ePNa is not available, the CheXpert model could provide the information from chest X-rays directly to clinicians, says Dean. “Using the CheXpert system, we found the interpretation time was very swift and the accuracy of the report to be very high,” he adds.
For the study, Intermountain radiologists categorized chest images from 461 Intermountain patients as being “likely,” “likely-uncertain,” “unlikely-uncertain,” or “unlikely” to have pneumonia. They also identified images they believed showed pneumonia in multiple parts of the lungs, and whether these patients had parapneumonic effusion, which is fluid build-up between the lungs and chest cavity.
The radiologists differed from each other in their categorizations in more than half of patients, as has been commonly shown in prior studies. The CheXpert model performance on the same images was comparable to the radiologists.
Researchers found that the CheXpert model outperformed the current system of using a radiologist to create radiology reports for all key pneumonia findings, plus NLP. It also did so in less than 10 seconds, compared to the 20 minutes to hours from NLP. NLP of radiology reports was the most frequent cause of errors within ePNa.
“A 2013 study published in JAMA Internal Medicine found that 59% of errors made by ePNA were due to NLP processing of radiologist reports, so we’re eager to replace it with a better, faster system,” Dean says.
Outside of ePNa concerns, emergency department physicians looking at radiology reports often are challenged to understand the unstructured language used by radiologists in interpreting shadows on chest x-rays, Dean adds.
The next step, he says, is for the CheXpert model to be used live in emergency departments, which he expects to happen in select Intermountain Healthcare hospitals this fall.