Article Content

The impact of deploying artificial intelligence (AI) for radiation cancer therapy in a real-world clinical setting has been tested by researchers at Princess Margaret Cancer Centre in a unique study involving physicians and their patients. A team of researchers directly compared physician evaluations of radiation treatments generated by an AI machine learning (ML) algorithm to conventional radiation treatments generated by humans. They found that, in the majority of the 100 patients studied, treatments generated using ML were deemed to be clinically acceptable for patient treatments by physicians.

  
Artificial Intellige... - Click to enlarge in new windowArtificial Intelligence. Artificial Intelligence

Overall, 89 percent of ML-generated treatments were considered clinically acceptable for treatments, and 72 percent were selected over human-generated treatments in head-to-head comparisons to conventional human-generated treatments.

 

Moreover, the ML radiation treatment process was faster than the conventional human-driven process by 60 percent, reducing the overall time from 118 hours to 47 hours. In the long term, this could represent a substantial cost savings through improved efficiency, while at the same time improving quality of clinical care, a rare win-win. The study also has broader implications for AI in medicine.

 

While the ML treatments were overwhelmingly preferred when evaluated outside the clinical environment, as is done in most scientific works, physician preferences for the ML-generated treatments changed when the chosen treatment, ML or human-generated, would be used to treat the patient. In that situation, the number of ML treatments selected for patient treatment was significantly reduced, issuing a note of caution for teams considering deploying inadequately validated AI systems.

 

Results by the study team led by Chris McIntosh, PhD, Leigh Conroy, PhD, Alejandro Berlin MD, MSc, and Tom Purdie, PhD, were published in Nature Medicine (2021; doi: 10.1038/s41591-021-01359-w).

 

"We have shown that AI can be better than human judgment for curative-intent radiation therapy treatment. In fact, it is amazing that it works so well," said McIntosh, scientist at the Peter Munk Cardiac Centre, Techna Institute, and Chair of Medical Imaging and AI at the Joint Department of Medical Imaging and University of Toronto. "A major finding is what happens when you actually deploy it in a clinical setting in comparison to a simulated one."

 

Added Purdie, who is a medical physicist at Princess Margaret Cancer Centre: "There has been a lot of excitement generated by AI in the lab, and the assumption is that those results will translate directly to a clinical setting. But we sound a cautionary alert in our research that they may not.

 

"Once you put ML-generated treatments in the hands of people who are relying upon it to make real clinical decisions about their patients, that preference towards ML may drop. There can be a disconnect between what's happening in a lab-type of setting and a clinical one." Purdie is Associate Professor in the Department of Radiation Oncology at the University of Toronto.

 

Study Details

In the study, treating radiation oncologists were asked to evaluate two different radiation treatments-either ML or human-generated ones-with the same standardized criteria in two groups of patients who were similar in demographics and disease characteristics.

 

The difference was that one group of patients had already received treatment so the comparison was a "simulated" exercise. The second group of patients were about to begin radiation therapy treatment, so if AI-generated treatments were judged to be superior and preferable to their human counterparts, they would be used in the actual treatments.

 

Oncologists were not aware of which radiation treatment was designed by a human or a machine. Human-generated treatments were created individually for each patient as per normal protocol by the specialized radiation therapist. In contrast, each ML treatment was developed by a computer algorithm trained on a high-quality, peer-reviewed database of radiation therapy plans from 99 patients previously treated for prostate cancer at Princess Margaret.

 

For each new patient, the ML algorithm automatically identifies the most similar patients in the database, using learned similarity metrics from thousands of features from patient images, and delineated target and healthy organs that are a standard part of the radiation therapy treatment process. The complete treatment for a new patient is inferred from the most similar patients in the database, according to the ML model.

 

Although ML-generated treatments were rated highly in both patient groups, the results in the pre-treatment group diverged from the post-treatment group. In the group of patients that had already received treatment, the number of ML-generated treatments selected over human ones was 83 percent. This dropped to 61 percent for those selected specifically for treatment prior to their treatment.

 

"In this study, we're saying researchers need to pay attention to a clinical setting," Purdie noted. "If physicians feel that patient care is at stake, then that may influence their judgment, even though the ML treatments are thoroughly evaluated and validated."

 

Conroy, a medical physicist at Princess Margaret, points out that following the highly successful study, ML-generated treatments are now used in treating the majority of prostate cancer patients at Princess Margaret.

 

That success is due to careful planning, judicious stepwise integration into the clinical environment, and involvement of many stakeholders throughout the process of establishing a robust ML program, she explained, adding that the program is constantly refined, oncologists are continuously consulted and give feedback, and the results of how well the ML treatments reflect clinical accuracy are shared with them.

 

"We were very systematic in how we integrated this into the clinic at Princess Margaret," said Berlin, clinician-scientist and radiation oncologist at Princess Margaret. "To build this novel software, it took about 6 months, but to get everyone on board and comfortable with the process, it took more than 2 years. Vision, audacity, and tenacity are key ingredients, and we are fortunate at Princess Margaret to have leaders across disciplines that embody these attributes."

 

The team is also expanding their work to other cancer sites, including lung and breast cancer, with the goal of reducing cardiotoxicity, a possible side effect of treatment.