- Research
- Open access
- Published:
Application of machine learning techniques for predicting survival in ovarian cancer
BMC Medical Informatics and Decision Making volume 22, Article number: 345 (2022)
Abstract
Background
Ovarian cancer is the fifth leading cause of mortality among women in the United States. Ovarian cancer is also known as forgotten cancer or silent disease. The survival of ovarian cancer patients depends on several factors, including the treatment process and the prognosis.
Methods
The ovarian cancer patients’ dataset is compiled from the Surveillance, Epidemiology, and End Results (SEER) database. With the help of a clinician, the dataset is curated, and the most relevant features are selected. Pearson’s second coefficient of skewness test is used to evaluate the skewness of the dataset. Pearson correlation coefficient is also used to investigate the associations between features. Statistical test is utilized to evaluate the significance of the features. Six Machine Learning (ML) models, including K-Nearest Neighbors , Support Vector Machine (SVM), Decision Tree (DT), Random Forest (RF), Adaptive Boosting (AdaBoost), and Extreme Gradient Boosting (XGBoost), are implemented for survival prediction in both classification and regression approaches. An interpretable method, Shapley Additive Explanations (SHAP), is applied to clarify the decision-making process and determine the importance of each feature in prediction. Additionally, DTs of the RF model are displayed to show how the model predicts the survival intervals.
Results
Our results show that RF (Accuracy = 88.72%, AUC = 82.38%) and XGBoost (Root Mean Squad Error (RMSE)) = 20.61%, R2 = 0.4667) have the best performance for classification and regression approaches, respectively. Furthermore, using the SHAP method along with extracted DTs of the RF model, the most important features in the dataset are identified. Histologic type ICD-O-3, chemotherapy recode, year of diagnosis, age at diagnosis, tumor stage, and grade are the most important determinant factors in survival prediction.
Conclusion
To the best of our knowledge, our study is the first study that develops various ML models to predict ovarian cancer patients’ survival on the SEER database in both classification and regression approaches. These ML algorithms also achieve more accurate results and outperform statistical methods. Furthermore, our study is the first study to use the SHAP method to increase confidence and transparency of the proposed models’ prediction for clinicians. Moreover, our developed models, as an automated auxiliary tool, can help clinicians to have a better understanding of the estimated survival as well as important features that affect survival.
Background
Cancer is the second deadliest disease around the world [1]. Due to the Coronavirus Disease 2019 (COVID-19) pandemic [2], cancer diagnosis and treatment were hindered [3]. With transforming and reorganizing healthcare systems to overcome COVID-19 difficulties, the screening, diagnosis, and treatment of cancers were not considered sufficiently. As a result, the pandemic raised serious concerns about the progression and increased mortality of cancers because clinicians do not have a tool for prioritizing high-risk patients in such a low-resource condition [4].
Ovarian cancer has a poor prognosis in most women since it is diagnosed at advanced stages [5]. This cancer is called forgotten cancer and is sometimes misdiagnosed [5]. Ovarian cancer has the fifth highest mortality rate among women living in the United States (US) [6]. The incidence rate of this cancer was 10.9 per 100,000 women in 2014–2018, and its mortality rate was 6.5 per 100,000 women in 2015–2019 [7]. It was estimated that new cases of ovarian cancer would be 21,410, accounting for 1.1% of all new cancer cases, and the estimated deaths due to ovarian cancer would be 13,770, accounting for 2.3% of all cancer deaths in the US in 2021 [4, 7].
Prognosis and survival prediction estimate the likelihood of recovery from a disease based on a patient’s clinical condition [8]. Determining a disease prognosis plays an important role, especially in malignant diseases such as cancer. It is one of the most important elements that help clinicians decide on more appropriate treatments. Survival prediction helps patients be informed about treatment decisions and reduce their anxiety [9].
Different methods have been used to predict cancer prognosis [10,11,12,13,14,15]. Some of these studies have paid attention to ovarian cancer, and researchers have utilized statistical methods to predict the survival of patients with ovarian cancer [11,12,13]. For instance, Stenzel et al. [12] analyzed the overall survival of ovarian cancer patients using a multivariable Cox proportional hazard model on the Surveillance, Epidemiology, and End Results (SEER) dataset. The results showed a 28% increased mortality risk in non-Hispanic black women compared to non-Hispanic white women. They also observed no difference in the risk of mortality between the survival of Hispanic women and non-Hispanic white women. Rutten et al. [13] predicted the five-year survival of ovarian cancer using a Cox proportional hazard model. Dataset was collected from three registries that included ovarian cancer patients who received bulking surgery. They also developed a nomogram to predict one-year, three-year, and five-year survival of ovarian cancer patients. The c-statistic their model achieved was 0.71. However, in statistical methods, all samples in datasets are not utilized, and the wrong sampling method could lead to data misinterpretation. Moreover, the statistical models are not explainable.
In addition to statistical techniques, the use of Machine Learning (ML) algorithms in the field of healthcare and medicine, to solve problems with different procedure and perspective, are growing dramatically [16]. Applying ML algorithms for predicting the survival of cancer patients is a relatively new field of study. Almost ML models are explainable models that use all samples of the training dataset and their output could be non-binary. Furthermore, they can be used for both classification and regression approaches simultaneously. Black box ML models are models that have low interpretability and transparency and the clinicians cannot see the models’ decision-making process. To address these challenges interpretable ML techniques have been developed and used to explain the process of predictions for black box ML models. Therefore, ML models can be made more understandable and reliable by using interpretable ML methods [17]. Shapley Additive Explanations (SHAP) is one of the methods based on game theory to interpret and explain the ML black box models. These models are used in various applications, including diagnosing, treating, and prognosis of different types of cancers. The literature shows that utilizing ML techniques in this domain has been promising, and there is some evidence that ML algorithms can outperform traditional statistical models [18,19,20,21,22,23,24].
In recent years, ovarian cancer has attracted researchers’ attention, and some studies developed ML classifiers to predict patients’ survival [14, 15]. For instance, Chen [14] used L2-regularized logistic regression to predict the mortality rate of fewer than 20 months of ovarian cancer patients in the SEER dataset. They achieved 0.62 for the Area Under the Curve (AUC) metric. Grimley et al. [15] used two datasets of ovarian carcinomas from the SEER database. The first dataset contained cases that had been staged for the extent of the tumor using T, N, and M criteria, and the second dataset was a derivative of the first one by treating age, histologic type, and grade as additional factors. They generated prognostic groups with the depiction in dendrograms using the Ensemble Algorithm for Clustering Cancer Data. Results revealed that the C-index of the International Federation of Gynecology and Obstetrics staging system was 0.7371, which is slightly lower than the C-index of 0.7391 from the Ensemble Algorithm for Clustering Cancer Data in the first dataset. The analysis of the second dataset revealed that the A and H could be smoothly integrated with the T, N, and M criteria. Survival data were classified into nine prognostic groups with a C-index of 0.7605. Nevertheless, our literature review showed that there are some research aspects that need further research. The identified research gaps in this field are as follows:
-
SEER database features have not been studied clinically to extract the most relevant features affecting the survival. In health informatics and bioinformatics fields of research, it is important that the parts of the study, including the dataset and its features, are clinically reliable.
-
In most studies, the dataset size is small and the number of samples used is low.
-
Even though the existing dataset is imbalanced, balancing techniques that can have a constructive impact on the performance of ML models have not been considered.
-
Statistical techniques, which are used to explain the relationship between variables, have been the main technique; and accurate prediction for survival has not been performed.
-
Classification is the common approach used for survival prediction, and there is no study utilizing a regression approach for predicting ovarian cancer patients’ survival.
-
Only binary classification was used; and multiple clinically meaningful classes have not been considered for different survival time intervals of ovarian cancer patients.
-
Interpretability and explainability of ML models in this field of research have not been addressed in previous studies.
In this study, six ML models of K-Nearest Neighbors (KNN), Support Vector Machine (SVM), Decision Tree (DT), Random Forest (RF), Adaptive Boosting (AdaBoost), and Extreme Gradient Boosting (XGBoost) were developed to predict the survival of ovarian cancer patients in two approaches of classification and regression [25,26,27]. KNN uses proximity to make predictions. It votes for the most frequent label among the nearest neighbors in classification and the average of the labels of the nearest neighbors in regression [26]. SVM uses a hyperplane to classify the data points. This algorithm aims to maximize the margin of dissociation between classes [26]. DT is an ML algorithm with a tree structure that consists of branches and nodes that illustrates every possible solution for a problem. This algorithm uses impurity metrics to make decisions [26, 27]. RF algorithm is an ensemble of multiple DTs, and the final outcome of RF is the aggregation of the DTs’ results [26]. AdaBoost is an ensemble of multiple classifiers or regressors. The final outcome of this algorithm is the combination of the results of its classifiers or regressors [26]. XGBoost consists of multiple weaker classifiers or regressors, the results of which are combined to determine the final output [25, 27]. It is the first time these algorithms are used to predict the survival of ovarian cancer patients based on the SEER database with classification and regression approaches simultaneously. These algorithms were modified since the best combination of hyperparameters was found using Grid Search. Grid Search is a method that evaluates all possible combinations of hyperparameters and picks the combination with the best results.
Objectives and contributions
The current study is designed to address the mentioned research gaps in this domain. Therefore, it aims to provide models in both classification and regression approaches to determine the survival time period classes and the number of survival months using ML algorithms. The proposed methods consist of a modified RF algorithm for the classification approach and a modified XGBoost algorithm for the regression approach. In summary, the followings are the main contributions of this study:
-
Feature engineering for SEER dataset is done based on discussion between engineers and clinicians of this study.
-
It is the first study that predicts ovarian cancer survival using both classification and regression approaches.
-
It is the first study that utilizes the mentioned six ML models to predict ovarian cancer patients’ survival based on the SEER dataset.
-
Classification and regression models of ML are used to accurately predict the number of survival months and the survival class of the patients, respectively.
-
Classes are defined based on clinical guidelines to have meaningful results from a clinical point of view.
-
Imbalanced data issue is addressed for the classification approach to enhance the performance of the models.
-
Models tuning are applied using Grid Search to find the best hyperparameters for the proposed ML models.
-
An ML interpretable method has been used to explain how models make decisions to increase the chance of deploying these models in real practice.
Materials and methods
The methodology of this study is illustrated in Fig. 1. Different components of our method are described in the following sections.
Study design and cohort selection
There are serious concerns about reporting the findings of utilizing ML in the health domain, and there are ongoing debates that most of the studies are not reproducible, and very difficult to judge their methodology and results. To address this issue, we followed transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) checklist [28, 29]. In this study, ovarian cancer data were collected from the SEER database. The SEER Stat program provides a database of cancer-related details and statistics. The SEER database is one of the most comprehensive and extensive population-based cancer registry data [4]. Ovarian cancer patients’ data between 2000 and 2016 were used in this study.
Outcome
For the classification approach, five clinically meaningful classes of survival namely, Class 0 (zero to 6 months), Class 1 (6 months to one year), Class 2 (one to three years), Class 3 (three to five years), Class 4 (five years and older) were defined. For the regression approach, the survival time in the number of months was considered.
Data preparation
SEER database is one of the most comprehensive and largest cancer databases that contains information on cancer patients of 48% of the US population [30]. Various information like data on tumor morphology and stage at diagnosis, primary tumor site, patient demographics, the first course of treatment, and follow-up for vital status (survival) from the 22 geographic areas of the US has been collected. It is supported by the Surveillance Research Program in National Cancer Institution’s Division of Cancer Control and Population Sciences [31].
For this study, the data of ovarian cancer patients between 2000 and 2016 was picked from the SEER database. This database contained many features; therefore, we performed feature selection. Feature selection is the process of removing irrelevant features and noise from the dataset to increase the accuracy and predictive power and decrease the learning time of the models. We performed feature selection under the supervision of a clinical researcher. So clinically significant and necessary features in predicting the survival of ovarian cancer patients were selected. The final dataset contained 42,827 samples and 17 features, as shown in Table 1. The survival months in this dataset were between 0 and 198 months.
Then, we used the Min-Max normalization method from the Sklearn library of Python. Doing so, the data was mapped to the range of (0,1) according to Eq. (1), where \({X}_{min}\) is the minimum value, and \({X}_{max}\) is the maximum value of a feature in the dataset.
The survival of ovarian cancer has been evaluated in terms of 1, 3, 5, and 10 years following the diagnosis of cancer [32, 33]. However, shorter intervals can be selected to make survival predictions more precise. Therefore, for the classification approach, five classes of survival months, namely Class 0 (zero to six months), Class 1 (six months to one year), Class 2 (one year to three years), Class 3 (three years to five years), and Class 4 (five years and above) were considered. Furthermore, to examine the survival of ovarian cancer patients accurately, patients’ survival in the number of months was also predicted using the regression approach.
Feature importance
The collected dataset included many numerical and categorical features. Based on discussions with clinicians, irrelevant features and incomplete records were removed from the dataset. Then, pearson correlation coefficient was used to investigate the associations between all features of the dataset. The correlation coefficient determines the degree and type of pairwise associations between features [34]. These coefficients have values between − 1 and 1, where 1 is the maximum correlation.
Determining the features’ importance of a dataset can be useful to support medical decisions and improve the patients’ quality of treatment [35]. It can be effective in predicting the survival of cancer patients to help clinicians make decisions by visualizing how decisions are made in the models [36, 37]. Therefore, we applied the SHAP library to interpret the model on our dataset. This vision and insight determine each feature’s importance and effectiveness in decision-making within the model [38].
Data imbalance
Data skewness measures the asymmetry of the distribution of a dataset. It is one of the inevitable challenges in many datasets, especially medical datasets. Skewed data can lead to a non-uniform sampling of the target feature and have an adverse effect on the performance of ML models. To solve this issue, Synthetic Minority Oversampling Technique (SMOTE) was used. SMOTE is a technique that creates synthetic data to oversample the minority classes in a dataset [39, 40]. After balancing the dataset, we have 14,778 samples for each class resulting in a total of 73,890 samples for all classes in the classification approach. One of the methods of testing the skewness of dataset is Pearson’s second coefficient of skewness test. This method determines the symmetry of the distribution of the dataset. Therefore, according to Eq. (2), this skewness test method was used to show the symmetry of the distribution of our dataset [41].
In Eq. (2), \(\overline{x}\) is the mean, \(m\) is the median, and \(s\) is the standard deviation.
Predictive models development
The KNN algorithm is one of the simplest supervised ML algorithms. This algorithm processes all the samples of training to make a prediction. For the classification approach, it finds the K-nearest neighbors and predicts the class with the majority of votes of the nearest neighbors. For the regression approach, this algorithm finds the K-nearest neighbors and predicts the desired value by calculating the average value of the nearest neighbors [26].
The SVM algorithm is an instance-based and supervised ML method that separates data samples using hyperplanes. It maximizes the margin between classes. The samples on one side of the line are considered similar and have the same category. This algorithm is used for both classification and regression approaches [26]. Equation (3) shows SVM’s optimization problem.
In Eq. (3), the tradeoff between guaranteeing that samples are located on the correct side of decision boundaries and expanding the size of decision boundaries is \(C\), and the number of samples is \(N\). To create non-linear boundaries, SVM uses kernels. The most common kernel is the Radial Basis Function kernel which is shown in Eq. (4).
In Eq. (4), the Euclidean distance of two vectors is \(\left\| {X - \mathop X\limits^{\prime } } \right\|^{2}\).
The DT algorithm is a widely used supervised ML algorithm that can be used in both classification and regression approaches. This algorithm determines which strategy is most likely to be successful using the Gini criterion, which is shown in Eq. (5). This algorithm consists of leaves, roots, and branches [26, 27].
In Eq. (5) ratio of class at the node of \(i\) is \(P\).
The RF algorithm is one of the most popular supervised ML algorithms. This algorithm is used for classification and regression approaches. To reach a more robust performance, it considers many trees instead of relying on one DT and makes predictions from each tree based on the majority vote. More trees in this algorithm often improve performance and prevent over-fitting. This algorithm is one of the most effective algorithms of ML in many applications [26].
The AdaBoost algorithm is a supervised ML technique that belongs to the boosting family of algorithms. In this algorithm, a problem is predicted by several different classifiers or regressors, usually DT, and the results are combined to determine the final result for the problem. The training of the classifiers or regressors of this algorithm is sequential, which means they are trained based on the result of the previous classifier or regressor [26].
One of the most powerful supervised ML algorithms is the XGBoost algorithm. This algorithm is an enhanced form of Gradient Boosting (GB). In this algorithm, several different classifiers or regressors predict a problem, and the combination of the results is the final result for the problem. Compared with GB, the model generalization capabilities of XGBoost are better due to the advanced regularization. It is also faster and more efficient than GB. This algorithm is used to predict classification and regression approaches [25, 27].
Moreover, five-fold cross-validation was applied to prevent over-fitting and evaluate the performance of implemented models [42]. In each iteration, 80% of the dataset was dedicated for training, and the rest part was used for testing.
This study determined a range of hyperparameters’ values for experimental implementation. They are shown in Table 10 in Appendix 1. Using the Grid Search method, the optimal hyperparameters of all models were achieved. The best hyperparameters selected by this method are shown in Tables 2 and 3 for the classification and regression approaches, respectively.
Predictive models evaluation
Various performance metrics were used for the evaluation of ML models. Ordinary Least Squares is a type of linear least-squares method to estimate the parameters in a linear regression model that describes the relationship between independent quantitative features and dependent features [43, 44]. It is used to show the statistically significant difference between the values of features. This test considers statistical significance (p-value) at different levels of 0.05, 0.01, and 0.001.
The confusion matrix is one of the classification evaluation criteria. This matrix is a square matrix whose dimension is \(n\times n\). The parameter n is equal to the number of classes in the classification [45]. After training and testing the models, we used the following ML evaluation criteria for the classification approach [46,47,48]. Accuracy is the ratio between the correct predictions of the data points to the total number of predictions. This criterion is used to evaluate the ML classification models [47]. The accuracy formula is shown in Eq. (6).
Precision or Positive Predictive Value (PPV) is one of the evaluation criteria of ML models. This criterion is the ratio between the total number of true positives to the total number of false positives and true positives. Precision indicates the accuracy of a model in a positive prediction [48]. Equation (7) shows the formula of Precision.
Negative predictive value (NPV) is calculated as the ratio between the total number of true negatives to the total number of false negatives and true negatives. It is one of the evaluation criteria of ML classification models that shows the probability that a person whose disease test is negative is truly healthy [48]. NPV formula is shown in Eq. (8).
Sensitivity or Recall is the ratio between the total number of true positives to the total number of false negatives and true positives. This criterion is used to evaluate ML models and the model’s ability in detecting positive samples [46]. The sensitivity formula is shown in Eq. (9).
Specificity is defined as the ratio between the total number of true negatives to the total number of false positives and true negatives. It is used in the performance evaluation of ML models. This criterion is important when the negative cases have priority. Because it shows the ability of the model to correctly detect true negatives. Equation (10) shows the formula of Specificity [46].
F1-Score calculates the harmonic mean of recall and precision and combines them into one metric so that models can be compared with one metric. A high F1-Score indicates low false positives and low false negatives [46, 48]. The F1-Score formula is shown in Eq. (11).
In Eqs. (6–10), TP stands for true positive, TN stands for true negative, FP stands for false positive, and FN stands for false negative.
The AUC means the area bounded by the Receiver Operating Characteristic curve. It is a measure of the overall accuracy of the models. The results of AUC range from 0 to 1, where a low value means bad model performance and a high value means an accurate model [47].
We also used the following evaluation criteria for the regression approach (prediction and prognosis of patients’ survival months). Root Mean Square Error (RMSE) calculates the standard deviation of the prediction errors. Prediction errors are the distance between the regression line and the data points. It measures the data’s concentration around the regression line. RMSE’s value is always non-negative, and a lower RMSE value is better than a higher value [49]. Equation (12) shows the formula of RMSE.
In Eq. (12), \(i\) is a sample, \(m\) means the number of samples, \({X}_{i}\) shows the actual target value for sample \(i\), and \({Y}_{i}\) denotes the predicted target value for sample \(i\).
R-Squared (R2) is a statistical measure that is used in the regression. It determines the proportion of variance in the dependent variable that is explained by the independent variable. R² indicates how well the regression model fits the data. Higher R² indicates that the relationship between the dependent variables and the regression model is strong [50]. The R² formula is shown in Eq. (13).
In Eq. (13), \(TSS\)means the total sum of squares and \(RSS\)means the residual sum of squares.
Results
Using the Pearson correlation coefficient, the correlation between features of the primary dataset was investigated. Some of these features had correlation coefficient values greater than 0.5 and consequently had a strong association with each other. The year of diagnosis was related to the insurance record with a correlation coefficient value of 0.85. Furthermore, the year of diagnosis was related to the CS site-specific factor 1 with a correlation coefficient value of 0.52. For other features, the correlation coefficient values were less than 0.5. Moreover, there was no complete inverse relevance between the features.
Furthermore, the skewness test was conducted in this study, where values greater than 1 and less than − 1 indicate high skewness. The initial value for the original dataset was 0.213, which is acceptable, but it can be improved by using a proper balancing technique. Therefore, SMOTE was applied to the original dataset, and the skewness was measured. The skewness value for the new dataset was − 0.032, which is closer to 0, indicating a more symmetric and balanced distribution.
Table 4 shows the comparison of the significance of the relation between the features of our dataset and the outcome variable. Some of the features were rejected at 0.05 and some at 0.01. The 0.001 level was used to ensure that there was no feature that we did not know had been rejected or accepted.
Classification models
Accuracy, PPV, NPV, sensitivity or recall, specificity, F1-Score, and AUC were calculated to evaluate the performance of the proposed classification models. Table 5 shows the average performance of the five folds of cross-validation of the proposed models.
Similar to other studies, accuracy and AUC metrics were considered as the most important criteria for evaluating the models [51, 52]. As shown in Table 5, on average, RF had the best performance for all evaluation criteria for the classification approach, which is bolded and italicized in Table 5. Furthermore, the computing time of one cycle of each fold of classification models is shown in Table 6. The lowest computing time is bolded in Table 6.
RF model achieved 88.72%, 71.91%, 92.96%, 71.81%, 92.95%, 71.78%, and 82.38%, on average, for all classes regarding the accuracy, PPV, NPV, sensitivity or recall, specificity, F1-Score, and AUC, respectively. More detailed values obtained for five folds of the RF model are illustrated in Table 11 of Appendix 1. Also, the RF model in the classification approach achieved the lowest execution time in one-fold among all other proposed models.
Moreover, the performances of all models for all folds were compared using confusion matrices, and the results were shown in Fig. 2. The fifth fold was used for visualizing the performance of the developed models based on the AUC metric, and the results are provided in Fig. 3. Moreover, as the best model (RF model), the confusion matrices and AUC diagrams of the RF model for all folds are shown in Fig. 7 and Fig. 8 in Appendix 1, respectively.
Regression models
To evaluate the performance of the proposed models of this study for the regression approach, the values of RMSE and R2 were calculated due to their importance which is in the same direction as the study’s criteria [53]. Furthermore, the computing time of one cycle of each fold of regression models is shown in Table 7. The lowest computing time is bolded in Table 7.
The average performance of the five folds cross-validation of the proposed models is shown in Table 8. As seen, XGBoost with RMSE = 20.61% and R2 = 0.4667 had the best performance for all evaluation metrics. The results of XGBoost for all five folds are listed in Table 12 in Appendix 1. Also, the XGBoost model in the regression approach achieved the lowest execution time in one-fold among all other proposed models.
Evaluation of the best-proposed models
The results showed that RF and XGBoost had the best performance for predicting the survival of ovarian cancer patients in classification and regression approaches, respectively.
The most important features determined by RF are shown in Fig. 4. As seen, the nine most important features with an average SHAP value greater than one are histologic type ICD-O-3, chemotherapy recode, year of diagnosis, age at diagnosis, summary stage, grade, marital status at diagnosis, laterality, and sequence number. The histologic type ICD-O-3 is the most discriminative feature for the purpose of predicting survival.
Figure 5 shows the effect of nine important features of this study and their impacts on the models’ output using the SHAP value. We randomly selected the 20th, 40th, 60th, 80th, and 100th trees created by our RF model with a depth of four to display predicted classes as shown in Figs. 9, 10, 11 and 12 in Appendix 1 and Fig. 6, respectively.
Discussion
In this study, we developed six ML models, including SVM, KNN, DT, RF, AdaBoost, and XGBoost, to predict the survival of ovarian cancer patients. For this aim, both classification and regression approaches for the SEER dataset were implemented. The promising results of this study are due to the appropriate size of the dataset’s samples and correct preprocessing. Therefore, we were able to get accurate outcomes with relatively few errors in the results [54].
The survival intervals of cancer patients are important and meaningful for clinicians, and they can plan patients’ treatment better based on that [55]. Therefore, in this study, with the help of an expert clinician, five classes including Class 0 (zero to 6 months), Class 1 (6 months to one year), Class 2 (one to three years), Class 3 (three to five years), Class 4 (five years and older) were considered. For the regression approach, the number of survival months was predicted.
Some studies have used the SEER dataset to predict ovarian cancer survival [11,12,13,14,15]; however, this field has some open areas that need further research. In the current study, we identified some of those gaps and addressed them. First, as interdisciplinary research, it is necessary to have a clinician in the research team to curate data, so our clinician evaluated the dataset carefully to identify the relevant features for survival prediction. Second, based on discussions with the clinician, we defined five classes that are meaningful from the clinical point of view. Third, to the best of our knowledge, the current approach in the studies is to develop classifiers; however, to have more accurate results, we implemented the regression approach to predict survival in months. Fourth, we paid attention to the interpretability and transparency of our system, so SHAP was used to explain the impact and importance of each feature on the model prediction performance. Moreover, we developed DTs of the RF technique that can help clinicians better understand the prediction process.
Our results showed that RF, with an accuracy of 88.72% and AUC of 82.38% on average, had the best performance for the classification approach, which is also better than models presented in the previous literature [14, 15]. For the regression approach, the XGBoost model with a RMSE of 20.61% and an R2 of 0.4667 had the best performance. Our proposed best models in each approach have the least computing time in one cycle run time compared to other proposed models.
Moreover, the performance of the ML models for each class was evaluated. Based on the results, for Class 0, the XGBoost model had the best performance, with an AUC of 84.34%. For Class 1, KNN obtained an AUC of 89.94%. For Class 2, RF model had an AUC of 71.01%. For Class 3, KNN gained an AUC of 83.08%. Finally, for Class 4, RF had an AUC of 89.10%. These findings show that the best ML algorithm in each survival class is different. The proposed RF and XGBoost models are both tree-based; therefore, they are explainable, which means they can be interpreted to make them understandable for humans, and their DTs can be easily displayed. These two attributes can help clinicians better comprehend the models’ decision-making. Furthermore, their computational cost is acceptable, and they can be easily used for large datasets.
Table 9 compares the dataset, method, outcome, performance, and explainability of two previous studies and this study. Our study and Chen’s study [14] used classification models, and Grimley et al. study [15] used a clustering model. As shown in the table, this study used a larger number of samples compared to the two other studies. Comparing the first study [14] with this study, it can be seen that the first study [14] used binary classification and predicted whether the patient would survive more than 22 months or not; however, we used multiclass classification to predict the survival intervals of patients. This range of classification provides more detailed information and planning possibilities for the clinicians. Despite using multiclass classification, the performance of our study is better than Chen’s study [14] in all criteria. Moreover, unlike Chen’s study [14], our study used an interpretation method to clarify the decision-making process of the proposed model.
The pandemic of COVID-19 has delayed the screening, diagnosis, and treatment of cancers, including ovarian cancer, which is expected to increase their mortality rates in future [56, 57]. Nevertheless, providing a tool that can accurately predict the survival of cancer patients will enable clinicians to recognize high-risk patients, prioritize them in case of using limited resources, and make evidence-based treatment decisions for them. The ML models proposed in this study have shown satisfactory performance in predicting the survival of ovarian cancer patients. The accuracy of predicting patients’ survival and diagnosis using ML models has increased significantly since 2000 [58]. In addition, the use of interpretable ML models has been able to show better and more understandable results than statistical models [59]. ML models’ interpretability and explainability, which show the effect of each feature on the prediction and decision-making of models, increase clinical and healthcare confidence in ML models [59]. In this study, histologic type ICD-O-3 is the most important feature in the model’s decision-making and has the highest SHAP value, as it is selected as the root of the tree in three of the five cases. This feature is a code that describes the morphology and topography of the tumors [60], both of which should be considered in predicting survival and planning treatment of ovarian cancer patients, as they influence survival [61, 62]; therefore, histologic type ICD-O-3 is an effective and important feature in ovarian cancer patients’ survival prediction. The effectiveness, importance, and usefulness of this feature have also been identified in other studies [62, 63].
It is noteworthy that ML models have the ability to predict from large and complex datasets, which highlights their increasing importance [37]. The prediction provided by ML algorithms is different from the epidemiological predictions since ML algorithms predict based on individual patients’ features and not on the base of a population average. Therefore, using ML algorithms as a valuable tool in times of crisis can be very helpful to clinicians, and the results of our study contribute to realizing the availability of such a tool using ML algorithms.
Strengths and limitations
This study has various strengths. The dataset features used were clinically meaningful and selected by our expert clinician. Correct and accurate preprocessing has been done on the dataset to avoid errors and mistakes during training and testing of ML models. For the first time, both classification and regression approaches simultaneously with multiple ML models have been implemented on the ovarian cancer patients’ dataset from the SEER database. To increase the performance of the models, the dataset has been balanced using the SMOTE method. The proposed models had an acceptable computational cost and were explainable. The SHAP method has been used to increase the confidence and clarity of clinicians in deciding on the best ML model in the classification approach. Moreover, the DTs of the best classification model of this study were drawn to provide better insight for clinicians.
However, this study has some limitations. First, it was not possible for us to implement the proposed models with more hyperparameters due to the limited resources. Second, it was not possible to validate our predictive models externally due to lack of similar available datasets.
Implication
The main audience of this study is clinical and ML researchers who are interested in the detailed analysis of the survival of cancer patients. Since we consider explainability, our best models can be tested as a practical tool to help clinicians to get an insight about the patients’ condition. Our results have many implications for managing ovarian cancer patients. These include updating or developing clinical guidelines and protocols based on the most important factors affecting these patients’ survival. Developing a clinical decision support tool based on our results is also another possible implication. The findings of this study can also be interesting for other researchers from different fields. Furthermore, developers in the field of ML can use the findings of this study to evaluate various techniques and create prediction models.
Conclusion
Ovarian cancer is one of the most common cancers in women. In this study, we developed ML techniques for both classification and regression approaches using the SEER dataset. An expert clinician helped us in preparing the data as well as design the classes to gain clinically meaningful results. To the best of our knowledge, our study is the first study using the regression approach for predicting ovarian cancer patients’ survival months.
In classification, RF, and in regression, XGBoost had the best performance. Both are tree-based and explainable. In addition, we considered the interpretability and transparency of the decision-making process by reporting the results using SHAP. The results of our study were promising and can be used as an auxiliary tool for clinicians to get insights into the condition of ovarian cancer patients, especially in situations like COVID-19, where a vast load goes to healthcare systems, and clinicians’ priorities would be changed.
We are currently collecting a new dataset of ovarian cancer patients in West Azerbaijan of Iran and will externally evaluate our proposed models in the future using this data.
Availability of data and materials
The dataset used in this study can be requested from the SEER source website at https://seerdataaccess.cancer.gov/seer-data-access.
Abbreviations
- COVID-19:
-
Coronavirus disease 2019
- ML:
-
Machine learning
- SEER:
-
Surveillance, Epidemiology, and End Results
- RF:
-
Random forest
- XGBoost:
-
Extreme gradient boosting
- US:
-
United States
- TRIPOD:
-
Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis
- SHAP:
-
Shapley additive explanations
- SMOTE:
-
Synthetic minority oversampling technique
- KNN:
-
K-nearest neighbors
- SVM:
-
Support vector machine
- DT:
-
Decision tree
- AdaBoost:
-
Adaptive boosting
- GB:
-
Gradient boosting
- PPV:
-
Positive predictive value
- NPV:
-
Negative predictive value
- AUC:
-
Area under curve
- RMSE:
-
Root mean square error
- R2 :
-
R-squared
References
Mlakar I, Lin S, Aleksandraviča I, Arcimoviča K, Eglītis J, Leja M, Salgado Barreira Á, Gómez JG, Salgado M, Mata JG, Batorek D. Patients-centered survivorShIp care plan after cancer treatments based on big data and artificial Intelligence technologies (PERSIST): a multicenter study protocol to evaluate efficacy of digital tools supporting cancer survivors. BMC Med Inf Decis Mak. 2021;21(1):1–4. https://doi.org/10.1186/s12911-021-01603-w.
Rikan SB, Azar AS, Ghafari A, Mohasefi JB, Pirnejad H. COVID-19 diagnosis from routine blood tests using Artificial Intelligence techniques. Biomed Signal Process Control. 2022;72:103263. https://doi.org/10.1016/j.bspc.2021.103263.
Al-Quteimat OM, Amer AM. The impact of the COVID-19 pandemic on cancer patients. American journal of clinical oncology. 2020.
Siegel RL, Miller KD, Fuchs HE, Jemal A. Cancer statistics. 2021. CA: a cancer journal for clinicians. 2021;71(1):7–33.
Tan JH, Sharpe L, Russell H. The impact of ovarian cancer on individuals and their caregivers: a qualitative analysis. Psycho-Oncol. 2021;30(2):212–20.
Ghoniem RM, Algarni AD, Refky B, Ewees AA. Multi-modal evolutionary deep learning model for ovarian cancer diagnosis. Symmetry. 2021;13(4):643.
Cancer of the ovary. - cancer stat facts [Internet resource]. SEER. [Accessed on 2022]. Available from: http://seer.cancer.gov/statfacts/html/ovary.html.
Hansebout RR, Cornacchi SD, Haines T, Goldsmith CH. How to use an article about prognosis. Can J Surg. 2009;52(4):328–36.
Turabian JL. Prognosis-based medicine-the importance of psychosocial factors: conceptualization from a case of acute pericarditis. Trends Gen Pract. 2018;1(1):1–2.
Department US. National cancer institute [Internet resource]. Cancer.gov. [Accessed on 2022 Jan]. Available from: https://surveillance.cancer.gov/publications/factsheets/SRP_Statistical%20Methods_Fact_Sheet.pdf.
Chen C, Markossian TW, Silva A, Tarasenko YN. Epithelial ovarian cancer mortality among Hispanic women: sub-ethnic disparities and survival trend across time: an analysis of SEER 1992–2013. Cancer epidemiology. 2018;52:134 – 41.
Stenzel AE, Buas MF, Moysich KB. Survival disparities among racial/ethnic groups of women with ovarian cancer: an update on data from the Surveillance, Epidemiology and End Results (SEER) registry. Cancer Epidemiol. 2019;62:101580.
Rutten MJ, Boldingh JH, Schuit E, Trum H, Van Driel W, Mol BW, Kenter GG, Buist MR. Development and internal validation of a prognostic model for survival after debulking surgery for epithelial ovarian cancer. Gynecologic Oncol. 2014;135(1):13–8.
Chen R. Machine learning for ovarian cancer: lasso regression-based predictive model of early mortality in patients with stage I and stage II ovarian cancer. medRxiv. 2020 Jan 1.
Grimley PM, Liu Z, Darcy KM, Hueman MT, Wang H, Sheng L, Henson DE, Chen D. A prognostic system for epithelial ovarian carcinomas using machine learning. Acta Obstetricia et Gynecologica Scandinavica. 2021.
Shastry KA, Sanjay HA. Machine learning for bioinformatics. In Statistical modelling and machine learning principles for bioinformatics techniques, tools, and applications 2020 (pp. 25–39). Springer, Singapore.
Petch J, Di S, Nelson W. Opening the black box: the promise and limitations of explainable machine learning in cardiology. Can J Cardiol. 2021. https://doi.org/10.1016/j.cjca.2021.09.004.
Elemento O, Leslie C, Lundin J, Tourassi G. Artificial intelligence in cancer research, diagnosis and therapy. Nat Rev Cancer. 2021;21(12):747–52.
Huang S, Yang J, Fong S, Zhao Q. Artificial intelligence in cancer diagnosis and prognosis: opportunities and challenges. Cancer lett. 2020;28:61–71.
Li J, Zhou Z, Dong J, Fu Y, Li Y, Luan Z, Peng X. Predicting breast cancer 5-year survival using machine learning: a systematic review. PloS One. 2021;16(4).
Lee C, Light A, Alaa A, Thurtle D, van der Schaar M, Gnanapragasam VJ. Application of a novel machine learning framework for predicting non-metastatic prostate cancer-specific mortality in men using the Surveillance, Epidemiology, and End Results (SEER) database. The Lancet Digital Health. 2021;3(3):e158-65.
Liu W, Wang S, Ye Z, Xu P, Xia X, Guo M. Prediction of lung metastases in thyroid cancer using machine learning based on SEER database. Cancer Med. 2022.
Doppalapudi S, Qiu RG, Badr Y. Lung cancer survival period prediction and understanding: deep learning approaches. Int J Med Inform. 2021;1:104371.
Senders JT, Staples P, Mehrtash A, Cote DJ, Taphoorn MJ, Reardon DA, Gormley WB, Smith TR, Broekman ML, Arnaout O. An online calculator for the prediction of survival in glioblastoma patients using classical statistics and machine learning. Neurosurgery. 2020;86(2):184–92.
Torlay L, Perrone-Bertolotti M, Thomas E, Baciu M. Machine learning–XGBoost analysis of language networks to classify patients with epilepsy. Brain Inf. 2017;4(3):159–69.
Shalev-Shwartz S, Ben-David S. Understanding machine learning: from theory to algorithms. Cambridge university press; 2014 p. 19.
Chen T, Guestrin C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining 2016 (pp. 785–794).
Cabitza F, Campagner A. The need to separate the wheat from the chaff in medical informatics: Introducing a comprehensive checklist for the (self)-assessment of medical AI studies. International Journal of Medical Informatics. 2021;153.
Collins GS, Reitsma JB, Altman DG, Moons KG. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD Statement. BMC Med. 2015;13(1):1–10.
SEER Program, National Cancer Institute (NCI). SEER Incidence Data, 1975–2017, Available:, 2019. http://seer.cancer.gov/data/.
National Cancer Institute, DCCPS, Surveillance Research Program: Surveillance, Epidemiology, and End Results (SEER) Program (www.seer.cancer.gov) SEER*Stat Database: Incidence - SEER 18 Regs Custom Data (with additional treatment fields), Nov 2018 Sub (1975–2016 varying) - Linked To County Attributes - Total US, 1969–2017 Counties. (2019). Based on the November 2018 submission.
Melamed A, Rauh-Hain JA, Gockley AA, Nitecki R, Ramirez PT, Hershman DL, Keating N, Wright JD. Association between overall survival and the tendency for cancer programs to administer neoadjuvant chemotherapy for patients with advanced ovarian cancer. JAMA Oncology. 2021;7(12):1782–90.
Gaitskell K, Hermon C, Barnes I, Pirie K, Floud S, Green J, Beral V, Reeves GK. Million Women Study Collaborators. Ovarian cancer survival by stage, histotype, and pre-diagnostic lifestyle factors, in the prospective UK Million Women Study. Cancer Epidemiology. 2022;76:102074. https://doi.org/10.1016/j.canep.2021.102074.
Sedgwick P. Pearson’s correlation coefficient. Bmj. 2012;345.
Rado O, Ali N, Sani HM, Idris A, Neagu D. Performance analysis of feature selection methods for classification of healthcare datasets. In intelligent computing-proceedings of the computing conference 2019 (pp. 929–938). Springer, Cham.
Laios A, Katsenou A, Tan YS, Johnson R, Otify M, Kaufmann A, Munot S, Thangavelu A, Hutson R, Broadhead T, Theophilou G. Feature selection is critical for 2-Year prognosis in advanced stage high grade serous ovarian cancer by using machine learning. Cancer Control. 2021;28:107.
Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Computational Struct Biotechnol J. 2015;1:8–17.
Kristjanpoller W, Michell K, Minutolo MC. A causal framework to determine the effectiveness of dynamic quarantine policy to mitigate COVID-19. Appl Soft Comput. 2021;1:107241.
Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: synthetic minority over-sampling technique. J Artif Intell Res. 2002;1:321–57.
Wang S, Dai Y, Shen J, Xuan J. Research on expansion and classification of imbalanced data based on SMOTE algorithm. Scientific reports. 2021;11(1):1–1.
Doane DP, Seward LE. Measuring skewness: a forgotten statistic?. Journal of statistics education. 2011;19(2).
Refaeilzadeh P, Tang L, Liu H. Cross-validation. Encyclopedia Database Syst. 2009;5:532–8.
Farahani HA, Rahiminezhad A, Same L. A comparison of partial least squares (PLS) and ordinary least squares (OLS) regressions in predicting of couples mental health based on their communicational patterns. Procedia-Soc Behav Sci. 2010;1:1459–63.
Judkins DR, Porter KE. Robustness of ordinary least squares in randomized clinical trials. Stat Med. 2016;35(11):1763–73.
Sidey-Gibbons JA, Sidey-Gibbons CJ. Machine learning in medicine: a practical introduction. BMC Med Res Methodol. 2019;19(1):1–8.
Parikh R, Mathai A, Parikh S, Chandra Sekhar G, Thomas R. Understanding and using sensitivity, specificity and predictive values. Indian J Ophthalmol. 2008;56(1):45–50.
Huang J, Ling CX. Using AUC and accuracy in evaluating learning algorithms. IEEE Trans Knowl Data Eng. 2005;17(3):299–310.
Wong HB, Lim GH. Measures of diagnostic accuracy: sensitivity, specificity, PPV and NPV. Proceedings of Singapore healthcare. 2011;20(4):316-8.
Chicco D, Warrens MJ, Jurman G. The coefficient of determination R-squared is more informative than SMAPE, MAE, MAPE, MSE and RMSE in regression analysis evaluation. PeerJ Comput Sci. 2021;5:e623.
Miles J. R-squared, adjusted R‐squared. Encyclopedia of statistics in behavioral science. 2005.
Zhang W, Ye B, Liang W, Ren Y. Preoperative prognostic nutritional index is a powerful predictor of prognosis in patients with stage III ovarian cancer. Sci Reports. 2017;7(1):1–8.
Cook NR. Statistical evaluation of prognostic versus diagnostic models: beyond the ROC curve. Clinical Chem. 2008;54(1):17–23.
Royston P, Moons KG, Altman DG, Vergouwe Y. Prognosis and prognostic research: developing a prognostic model. Bmj. 2009;338.
Riley RD, Ensor J, Snell KI, Harrell FE, Martin GP, Reitsma JB, Moons KG, Collins G, Van Smeden M. Calculating the sample size required for developing a clinical prediction model. Bmj. 2020;368.
Mackillop WJ. The importance of prognosis in cancer medicine. TNM Online. 2003.
Blay JY, Boucher S, Le Vu B, Cropet C, Chabaud S, Perol D, Barranger E, Campone M, Conroy T, Coutant C, De Crevoisier R. Delayed care for patients with newly diagnosed cancer due to COVID-19 and estimated impact on cancer mortality in France. ESMO Open. 2021;6(3).
Jacome LS, Deshmukh SK, Thulasiraman P, Holliday NP, Singh S. Impact of COVID-19 pandemic on ovarian cancer management: adjusting to the new normal. Cancer Manage Res. 2021;13:359.
Cruz JA, Wishart DS. Applications of machine learning in cancer prediction and prognosis. Cancer Inform. 2006;2:117693510600200030.
Moncada-Torres A, van Maaren MC, Hendriks MP, Siesling S, Geleijnse G. Explainable machine learning can outperform cox regression predictions and provide insights in breast cancer survival. Sci Rep. 2021;26(1):1–3.
Quah S. International encyclopedia of public health. 6: Academic Press; 2016. pp. 304–11). Elsevier.
Matz M, Coleman MP, Sant M, Chirlaque MD, Visser O, Gore M, Allemani C, Bouzbid S, Hamdi-Chérif M, Zaidi Z, Bah E. The histology of ovarian cancer: worldwide distribution and implications for international survival comparisons (CONCORD-2). Gynecol Oncol. 2017;144(2):405–13.
World Health Organization. International classification of diseases for oncology (ICD-O)–3rd edition, 1st revision.
Tercek A, Galbo A, Makhani S, Bouz A, Chung-Bridges K. Racial disparities in women with serous epithelial ovarian cancer: a Surveillance, Epidemiology, and end results (SEER) survival analysis. Gynecol Oncol. 2021;1;162:251–2.
Acknowledgements
We would like to thank the Machine Learning Lab of Urmia University for providing the resources needed to implement this study.
Funding
There is no funding attached to this study.
Author information
Authors and Affiliations
Contributions
ASA, SBR, AN, and JBM were involved in the conception and design of this study. ASA, SBR, MBM, and HP prepared the dataset, and ASA and SBR performed the analysis. ASA, SBR, AN, JBM, HP, and UKW interpreted the results. ASA and SBR drafted the manuscript, and all authors (ASA, SBR, AM, JBM, HP, MBM, UKW) contributed to writing the final draft and prepared the final manuscript. All authors read and approved by the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors report no conflict of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Amir Sorayaie Azar and Samin Babaei Rikan equally contributed to this work as first co-authors.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Sorayaie Azar, A., Babaei Rikan, S., Naemi, A. et al. Application of machine learning techniques for predicting survival in ovarian cancer. BMC Med Inform Decis Mak 22, 345 (2022). https://doi.org/10.1186/s12911-022-02087-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12911-022-02087-y