An increasing volume of prostate biopsies and a worldwide shortage of urological pathologists puts a strain on pathology departments. Additionally, the high intra-observer and inter-observer variability in grading can result in overtreatment and undertreatment of prostate cancer. To alleviate these problems, we aimed to develop an artificial intelligence (AI) system with clinically acceptable accuracy for prostate cancer detection, localisation, and Gleason grading.
We digitised 6682 slides from needle core biopsies from 976 randomly selected participants aged 50–69 in the Swedish prospective and population-based STHLM3 diagnostic study done between May 28, 2012, and Dec 30, 2014 (ISRCTN84445406), and another 271 from 93 men from outside the study. The resulting images were used to train deep neural networks for assessment of prostate biopsies. The networks were evaluated by predicting the presence, extent, and Gleason grade of malignant tissue for an independent test dataset comprising 1631 biopsies from 246 men from STHLM3 and an external validation dataset of 330 biopsies from 73 men. We also evaluated grading performance on 87 biopsies individually graded by 23 experienced urological pathologists from the International Society of Urological Pathology. We assessed discriminatory performance by receiver operating characteristics and tumour extent predictions by correlating predicted cancer length against measurements by the reporting pathologist. We quantified the concordance between grades assigned by the AI system and the expert urological pathologists using Cohen’s kappa.
The AI achieved an area under the receiver operating characteristics curve of 0·997 (95% CI 0·994–0·999) for distinguishing between benign (n=910) and malignant (n=721) biopsy cores on the independent test dataset and 0·986 (0·972–0·996) on the external validation dataset (benign n=108, malignant n=222). The correlation between cancer length predicted by the AI and assigned by the reporting pathologist was 0·96 (95% CI 0·95–0·97) for the independent test dataset and 0·87 (0·84–0·90) for the external validation dataset. For assigning Gleason grades, the AI achieved a mean pairwise kappa of 0·62, which was within the range of the corresponding values for the expert pathologists (0·60–0·73).
An AI system can be trained to detect and grade cancer in prostate needle biopsy samples at a ranking comparable to that of international experts in prostate pathology. Clinical application could reduce pathology workload by reducing the assessment of benign biopsies and by automating the task of measuring cancer length in positive biopsy cores. An AI system with expert-level grading performance might contribute a second opinion, aid in standardising grading, and provide pathology expertise in parts of the world where it does not exist.
Swedish Research Council, Swedish Cancer Society, Swedish eScience Research Center, EIT Health.
In order to determine whether someone has prostate cancer or not, a doctor will take a prostate biopsy. This biopsy is read under microscopic examination by the doctor to differentiate harmless from malignant tissue. Not only does the doctor base her/his diagnosis on this tissue, he/she also uses it to develop a treatment plan. However, logical human inaccuracies originating from a simple situation such as one doctor drawing different conclusions from observing the same material more than once may lead to over- or undertreatment of prostate cancer.
In this study, for the first time ever, the authors evaluated a tool which we only knew from science fiction stories until recently: Artificial Intelligence (AI). The AI-assisted prostate evaluation serves the same function as the doctor’s microscopic examination. The authors found that the AI system performed similarly to experienced doctors in detecting prostate cancer and grading the malignant tissue. They concluded that the implementation of such an AI system in the daily routine could reduce the number of missed cancer diagnoses and could decrease the doctor’s workload.