Next Article in Journal
Decursinol Angelate Inhibits Glutamate Dehydrogenase 1 Activity and Induces Intrinsic Apoptosis in MDR-CRC Cells
Previous Article in Journal
Preoperative CT-Based Skeletal Muscle Mass Depletion and Outcomes after Total Laryngectomy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Cancer Recurrence Using Systemic Inflammatory Markers and Machine Learning after Concurrent Chemoradiotherapy for Head and Neck Cancers

1
Department of Otorhinolaryngology-Head & Neck Surgery, Inje University College of Medicine, Ilsan Paik Hospital, Goyang-Si 10380, Republic of Korea
2
Medical AI Research Center, Samsung Medical Center, Seoul 06351, Republic of Korea
3
Department of Data Convergence and Future Medicine, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea
4
Department of Otolaryngology-Head & Neck Surgery, Sungkyunkwan University School of Medicine, Samsung Medical Center, Seoul 06351, Republic of Korea
5
Center for Thyroid Cancer, Department of Otolaryngology-Head and Neck Surgery, Research Institute and Hospital, National Cancer Center, Goyang-si 10408, Republic of Korea
6
Divison of Hematology and Medical Oncology, Department of Medicine, Sungkyunkwan University School of Medicine, Samsung Medical Center, Seoul 06351, Republic of Korea
7
Department of Radiation Oncology, Sungkyunkwan University School of Medicine, Samsung Medical Center, Seoul 06351, Republic of Korea
8
Department of Biomedical Informatics, CHA University School of Medicine, CHA University, Seongnam-Si 13488, Republic of Korea
9
Institute of Biomedical Informatics, School of Medicine, CHA University, Seongnam-Si 13488, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2023, 15(14), 3540; https://doi.org/10.3390/cancers15143540
Submission received: 1 June 2023 / Revised: 3 July 2023 / Accepted: 6 July 2023 / Published: 8 July 2023

Abstract

:

Simple Summary

The neutrophil-to-lymphocyte ratio (NLR) and the platelet-to-lymphocyte ratio (PLR) are known as prognosticators in head and neck cancers, but the clinical significance of those values after treatment remains unclear. In this study, we aimed to investigate the relationship between temporal changes in the NLR and PLR values after treatment and cancer recurrence in head and neck cancer patients. We proposed developing a DNN model that utilizes these changes to predict recurrence after CCRT, comparing its performance with traditional machine learning (ML) models. The results of this study showed that the DNN model outperformed the other models, and feature selection using the recursive feature elimination algorithm further improved the performance, with ROC-AUC and PR-AUC values of 0.883 ± 0.027 and 0.778 ± 0.042, respectively. The results could improve prognostication and risk assessment in head and neck cancers. Additionally, this study may inspire further research and advancements in utilizing advanced ML techniques for personalized cancer management and surveillance.

Abstract

Pretreatment values of the neutrophil-to-lymphocyte ratio (NLR) and the platelet-to-lymphocyte ratio (PLR) are well-established prognosticators in various cancers, including head and neck cancers. However, there are no studies on whether temporal changes in the NLR and PLR values after treatment are related to the development of recurrence. Therefore, in this study, we aimed to develop a deep neural network (DNN) model to discern cancer recurrence from temporal NLR and PLR values during follow-up after concurrent chemoradiotherapy (CCRT) and to evaluate the model’s performance compared with conventional machine learning (ML) models. Along with conventional ML models such as logistic regression (LR), random forest (RF), and gradient boosting (GB), the DNN model to discern recurrences was trained using a dataset of 778 consecutive patients with primary head and neck cancers who received CCRT. There were 16 input features used, including 12 laboratory values related to the NLR and the PLR. Along with the original training dataset (N = 778), data were augmented to split the training dataset (N = 900). The model performance was measured using ROC-AUC and PR-AUC values. External validation was performed using a dataset of 173 patients from an unrelated external institution. The ROC-AUC and PR-AUC values of the DNN model were 0.828 ± 0.032 and 0.663 ± 0.069, respectively, in the original training dataset, which were higher than the ROC-AUC and PR-AUC values of the LR, RF, and GB models in the original training dataset. With the recursive feature elimination (RFE) algorithm, five input features were selected. The ROC-AUC and PR-AUC values of the DNN-RFE model were higher than those of the original DNN model (0.883 ± 0.027 and 0.778 ± 0.042, respectively). The ROC-AUC and PR-AUC values of the DNN-RFE model trained with a split dataset were 0.889 ± 0.032 and 0.771 ± 0.044, respectively. In the external validation, the ROC-AUC values of the DNN-RFE model trained with the original dataset and the same model trained with the split dataset were 0.710 and 0.784, respectively. The DNN model with feature selection using the RFE algorithm showed the best performance among the ML models to discern a recurrence after CCRT in patients with head and neck cancers. Data augmentation by splitting training data was helpful for model performance. The performance of the DNN-RFE model was also validated with an external dataset.

1. Introduction

Systemic inflammation is known to suppress the anticancer immune response [1,2], and systemic inflammatory markers have been associated with poor prognosis for various cancers [3,4,5,6,7,8,9,10]. Neutrophilia and thrombocytosis have been shown to reflect tumor progression. In addition, lymphopenia, which indicates a compromised antitumor immune response, has been associated with a poor prognosis in various advanced solid tumors. The neutrophil-to-lymphocyte ratio (NLR) and the platelet-to-lymphocyte ratio (PLR) may capture the combined effects of these opposing factors. High NLR and PLR values have been widely recognized as poor prognostic indicators for recurrence and survival in head and neck cancers [3,11,12,13,14]. However, it has not been studied whether temporal changes in the NLR and PLR values after treatment are related to the development of recurrence. Post treatment, patients are followed up continuously for cancer recurrence, and blood tests are performed along with imaging studies. Resultantly, large amounts of laboratory data have been accumulated. Apart from the pretreatment values of the NLR and PLR, the temporally collected values of NLR and PLR after the end of the treatment may have separate clinical significance. They can be used to detect recurrence during the follow-up of cancer patients, and, considering the vast amount of the data, they are also suitable for use in machine learning (ML) or deep learning (DL).
ML is a collection of methods that allow software applications to develop algorithms to predict outcomes from existing datasets without being explicitly programmed. DL is an advanced subset of machine learning based on artificial neural networks with representation learning. Deep neural network (DNNs), which are representative architectures of DL, have recently made significant progress in developing predictive models in the medical field using electronic medical record (EMR) datasets [15,16,17]. It is known that DNNs are superior to conventional ML models in handling vast amounts of data.
In this study, we aimed to develop a DNN model to discern cancer recurrence from the temporal NLR and PLR values during follow-up after chemoradiation and to evaluate the performance of the DNN model compared with conventional ML models.

2. Materials and Methods

2.1. Patients

This study was conducted on a cohort of primary head and neck cancer patients who had received definitive concurrent chemoradiotherapy (CCRT) from March 1994 to March 2019. Initially, 1212 consecutive patients were screened for enrollment, and among them, patients were excluded with other diseases that could affect blood test values, for example, synchronous or metachronous malignancies, hematologic diseases such as idiopathic thrombocytopenic purpura, HIV infection, liver cirrhosis, and active tuberculosis. We also excluded patients with missing analytical items, including missing blood test data, missing medical records on recurrences, and missing follow-up status, as well as patients who smoked. Patients were excluded from this study who developed recurrence within three months after completion of CCRT, which was considered to be a remnant disease. Lastly, those who were lost to follow-up within three months were also excluded. Finally, 758 patients were included. This study was approved by the Institutional Review Board of Samsung Medical Center (SMC IRB file no. 2020-06-125-001) and was conducted following the Helsinki Declaration. Informed consent was waived because of the retrospective design.

2.2. Training Dataset and Input Features

This study used the EMR data of 758 primary head and neck cancer patients who had received definitive concurrent chemoradiotherapy (CCRT) from March 1994 to March 2019. The original dataset was retrieved via a retrospective review of the patients’ EMRs. The outcome label is a binary variable indicating the development of recurrence. Recurrence was decided based on clinical, radiological, or pathological descriptions from EMRs and included regional, local, and distant recurrences. We utilized 16 input features for training the ML models, which included each patient’s age, sex, smoking status, stage (based on the AJCC 7th system), and 12 blood laboratory features. Laboratory values up to the diagnosis of recurrence were used for recurrent cases, and laboratory values up to the end of the follow-up with a 5-year limit were used for non-recurrent cases. Recent deep learning technologies have the capability to capture the temporal patterns of time series data; however, it is often very difficult to learn such a pattern if the amount of data is limited and data have been acquired irregularly for each patient. In addition, the duration of follow-ups differed from each other, and the visits were also often irregular. Therefore, we extracted 12 features from the irregular longitudinal blood laboratory values for each case (Table 1) [18]. The NLR and PLR values were calculated by dividing the neutrophil count and the platelet count by the lymphocyte count, respectively. For the NLR, the following six blood laboratory features were calculated during the corresponding period in recurrent patients: mean NLR (NLR mean), maximum–minimum NLR (NLR max–min), NLR after treatment (NLR Tx), NLR at the time of recurrence (NLR R), the difference between NLR R and NLR mean (NLR R-mean), and the difference between NLR R and NLR Tx (NLR R-Tx). The NLR Tx and PLR Tx were obtained from 14 days to 90 days after cessation of CCRT. The NLR R and PLR R were obtained from 90 days before the detection of recurrence to 30 days after the detection of recurrence. In non-recurrent patients, the NLR R means the NLR at the end of the period. For the PLR, six blood test indices were extracted in the same manner.
Missing values were removed, and categorical variables were encoded using the one-hot encoding technique since there was no ordinal relationship. Subsequently, the 12 extracted input variables were normalized to a standard scale. Additionally, recursive feature elimination was performed to assess the impact of each input variable on the response variable. The contribution of each variable to the prediction was calculated as a discrete interval encompassing all the variables in the training dataset. The response probability was estimated for each range of the input variable, while keeping the other variables at their average value for numeric variables or at zero for binary variables. In each iteration, the input variable with the smallest dynamic range of the output response was successively removed until only one variable remained.

2.3. Split Training Dataset

In order to overcome the limited size of the training data, we tried to adopt data augmentation by data splitting. For each recurrent case, the temporal blood laboratory values were split into two groups based on 180 days before the diagnosis of recurrence (Figure 1a). Values more than 180 days before recurrence were reclassified as the values of the non-recurrent state. We hypothesized that data until 180 days before recurrence could be treated as negative (non-recurrent) samples, which could be helpful for better model performance. In that way, a modified dataset was also built (Figure 1b).

2.4. Machine Learning Models

For both the original and the modified dataset, traditional ML models and a DNN model were built. Among the traditional supervised ML algorithms, random forest (RF), logistic regression (LR), and gradient boosting (GB) were used. On the one hand, logistic regression is a statistical model used for binary classification problems. It estimates the probability of an instance belonging to a particular class by utilizing a logistic function. This model assumes a linear relationship between the input features and the log-odds of the target class; however, it may struggle to capture complex nonlinear relationships and may not perform as effectively as more advanced models in highly complex tasks.
Random forest, on the other hand, is an ensemble learning method that combines multiple decision trees to make predictions. It creates a forest of decision trees, where each tree is trained on a random subset of the data and features. Random forest is capable of performing both classification and regression tasks and can handle complex interactions and nonlinear relationships in the data; however, it can be prone to overfitting if the number of trees is too high and may not perform as well on high-dimensional datasets.
Gradient boosting is another ensemble learning technique that builds an ensemble of weak prediction models, typically decision trees, in a sequential manner. Gradient boosting combines the predictions of multiple weak models to make a final prediction. It is capable of handling complex relationships, capturing interactions, and performing well on structured and tabular data; however, it can be computationally expensive and requires careful tuning of hyperparameters.
In contrast to logistic regression, which assumes a linear relationship between the input features and the log-odds of the target class, deep neural networks (DNNs) can approximate any complex function by learning hierarchical representations through hidden layers. This flexibility enables DNNs to handle highly nonlinear data and to capture intricate interactions and dependencies between features. DNNs excel in tasks where the relationships in the data are complex and may not be easily captured by simpler models.
To develop the ML models, the entire dataset was divided into 10 folds by applying stratified random splits. The model was trained with nine of the folds, and the performance of the model was measured repeatedly with the remaining fold. The performance of the model was measured using a receiver operating curve and the area under the curve (ROC-AUC) and precision recall and the area under the curve (PR-AUC) values. The models were developed on Ubuntu Environment using Python 3.5. The DNN algorithms were developed, including all input features or selected features using the RFE method [19]. In the DNN-RFE model, the weakest features were removed until the ROC-AUC and PR-AUC values reached their peaks. The hyperparameters of the DNN models were tuned to reach the highest mean of the ROC-AUC and PR-AUC values. The number of units, batch size, learning rate, and dropout rate were tuned. Adam was used as an optimizer, and the optimal learning rate was 0.001 [20].

2.5. External Validation Dataset

The validation cohort was enrolled separately in an unrelated institution. A total of 173 primary head and neck cancer patients who had received definitive concurrent chemoradiotherapy (CCRT) from April 2002 to March 2019 were enrolled for validation. This validation dataset was also built in two versions, i.e., the original validation dataset and a modified validation dataset (split dataset).

3. Results

3.1. Patients’ Characteristics

The clinicopathological profiles of 778 patients included in the training dataset study are presented in Table 2. The mean age of the patients was 60.4 years (standard deviation, 13.0 years). There were 625 males and 153 females. The two most common tumor sites were the nasopharynx (38.0%) and the oropharynx (29.3%). Pathologically, squamous cell carcinoma (64.7%) and undifferentiated carcinoma (31.2%) accounted for most cases. Approximately 90% of patients were in advanced stages. In total, there were 489 cases (62.9%) in stage IV and 209 cases (26.9%) in stage III. Among all enrolled cases, 206 (26.5%) cases had a recurrence, 108 cases had locoregional recurrences without distant metastasis, and 98 cases had distant metastasis with or without locoregional recurrence. The five-year recurrence-free survival (RFS) rate was 66.5%, and the median RFS was 189.8 months.
With data splitting of the recurrent cases, which had no recurrence within 180 days after the end of the treatment, 117 non-recurrent cases were newly created (Figure 1). Resultantly, 206 recurrent cases and 689 non-recurrent cases were included in the split training dataset.
In the external validation dataset, 73 recurrent cases and 100 non-recurrent cases were included. After data splitting of recurrent cases, 73 recurrent cases and 152 non-recurrent cases were included in the split validation dataset. We tried to split the external validation dataset in order to test whether our ML models could accurately recognize the sample data, until 6 months before recurrence, as negative (non-recurrent) cases. The clinicopathological characteristics of the external validation dataset are presented in Table S1.

3.2. Cross-Validation Performance on Training Datasets

The ROC-AUC values of the LR, RF, and GB models were 0.696 ± 0.059, 0.812 ± 0.049, and 0.600 ± 0.052, respectively, on the original (non-split) training dataset (Figure 2a). The ROC-AUC value of the DNN model using 16 input features was 0.828 ± 0.032 on the original training dataset (Table 3, Figure 2a). The PR-AUC values of the LR, RF, GB, and DNN models were 0.454 ± 0.079, 0.622 ± 0.081, 0.346 ± 0.055, and 0.663 ± 0.069, respectively, on the original training dataset.
When the models were trained using a split training dataset, the cross-validation performances were also measured. On the split training dataset, the ROC-AUC values of the LR, RF, GB, and DNN models were 0.700 ± 0.053, 0.751 ± 0.053, 0.735 ± 0.057, and 0.837 ± 0.038, respectively (Figure 2b). The PR-AUC values of the LR, RF, GB, and DNN models were 0.456 ± 0.072, 0.539 ± 0.084, 0.516 ± 0.085, and 0.666 ± 0.045, respectively.
When the RFE algorithm was applied to the DNN model on the original training dataset, the top three features which maximized the ROC-AUC and the PR-AUC values of the DNN model were selected (Figure 3a). The selected features were PLR mean, PLR Tx, and PLR max–min. With these three features, the ROC-AUC and the PR-AUC values of the DNN-RFE model were 0.884 ± 0.032 and 0.788 ± 0.032, respectively, on the original training dataset (Table 4). The DNN-RFE algorithm was applied to the split training dataset, and PLR R, PLR R-Tx, PLR mean, PLR Tx, and PLR max–min features were selected (Figure 3b). With these top five features, the ROC-AUC and the PR-AUC values of the DNN-RFE model were 0.889 ± 0.032 and 0.771 ± 0.044, respectively, on the split training dataset. When these top five features were used on the original training dataset, the ROC-AUC and the PR-AUC values were 0.883 ± 0.027 and 0.778 ± 0.042, respectively (Table 4).
The DNN-RFE model trained with a split training dataset had the maximal performance, and five blood laboratory features were included in the model (PLR R, PLR R-Tx, PLR mean, PLR Tx, and PLR max–min).

3.3. External Validation (Figure 4)

We evaluated our ML models on the original external dataset (N = 173). Each ML and DNN model trained with the original and the split training dataset was evaluated on the original external validation dataset (N = 173) (Figure 4). From the original training dataset, the ROC-AUC values of the LR, RF, GB, DNN, and DNN-RFE models were 0.572, 0.748, 0.585, 0.643, and 0.710, respectively (fourth column in Table 5). When we evaluated the models trained with a split training dataset, the ROC-AUC values of the LR, RF, GB, DNN, and DNN-RFE models were 0.556, 0.531, 0.672, 0.766, and 0.784, respectively (third column in Table 5).
Figure 4. Overview of the external validation process for both original dataset model and split dataset model on each validation dataset.
Figure 4. Overview of the external validation process for both original dataset model and split dataset model on each validation dataset.
Cancers 15 03540 g004
Additionally, we evaluated our ML models on a split external validation dataset again (N = 225) (Figure 4b). When we evaluated the split validation dataset, the ML and DNN models trained with the split training dataset showed higher performances than the models trained with the original training dataset. The ROC-AUC value of the DNN-RFE model trained with the split training dataset was higher than that of the DNN-RFE model trained with the original dataset (0.845 vs. 0.708 in Table 5).

4. Discussion

Inflammatory responses within the tumor microenvironment play important roles in cancer initiation and progression [21,22,23]. The systemic inflammatory response is associated with peri-tumoral or intra-tumoral inflammation. The systemic inflammatory response can suppress the antitumor activity of immune cells, such as activated T cells and natural killer cells [1,2]. Recently, systemic inflammatory markers have been widely studied as prognostic factors in various cancers, including head and neck cancers. Among several markers, the NLR and the PLR values are the most well-established prognosticators. A meta-analysis of 100 clinical studies covering a variety of solid tumors, including head and neck cancers, showed that higher NLR values were significantly associated with shorter OS with a hazard ratio (HR) of 1.81 (95% CI = 1.67 to 1.97). Higher NLR values were also associated with shorter disease-free survival (DFS) (HR, 2.27, 95% CI = 1.67 to 1.97) [24]. An elevated PLR has also been reported to be a poor prognostic factor. In a meta-analysis of 26 studies including 13,964 patients, elevated PLR values were a significant factor for worse OS with a HR of 1.60 (95% CI = 1.35–1.90) [25]. Various inflammatory markers have also been studied for head and neck cancers alone, and NLR and PLR values have been reported as independent prognostic factors [12,13,14,26].
Post-treatment NLR and PLR values have also been reported to have prognostic value in various cancers. Radiation therapy (RT) and concurrent chemoradiation therapy (CCRT) for head and neck cancers can frequently induce lymphopenia and other blood cell count changes [27,28]. The resultant increases in NLR or PLR values can have prognostic significance independent of the pretreatment NLR or PLR values [29,30,31].
However, studies on the NLR and the PLR have focused on the prognostic significance of either pretreatment or post-treatment values. Previous studies have been intended to evaluate the risk of future death or cancer recurrence in patients with presently elevated NLR or PLR values. Therefore, pretreatment or post-treatment NLR and PLR values have been classified as prognostic factors such as tumor stage. Prominent prognostic factors for cancer recurrence include stage, pathology, and site. Tumors diagnosed at an advanced stage, those exhibiting aggressive pathology with poor differentiation, or those located in sites such as the oral tongue are associated with poor prognosis. In addition to these factors, patients with higher NLR or PLR values at diagnosis are now believed to have an increased likelihood of experiencing recurrence after treatment.
Unlike previous studies, we employed serially measured NLR and PLR values after the completion of treatment, rather than point values at diagnosis. Our objective was not to assess the prognostic values of those systemic inflammatory markers. Instead, we aimed to determine whether the temporal values measured after treatment could indicate the presence of recurrence, if it occurs. Various modalities, such as sonography or CT scan, can detect cancer recurrence during the follow-up after treatment. Systemic inflammatory markers and blood test results can be used as intermediate means to detect recurrence, although not as directly as radiologic tests. As we used in this study, blood tests are continuously checked after treatment, and enormous amounts of test data can be accumulated. Because an increase in systemic inflammatory markers can reflect progression of cancers, we hypothesized that temporal changes in NLR and PLR values can be used for a prediction model to detect cancer recurrence.
The prediction of cancer recurrence and the identification of risk factors have mainly been performed using logistic regression. With advances in ML techniques, many ML models for predicting cancer recurrence have recently been reported [32,33,34]. DL has been shown to enhance the performance and generalizability of ML models. DL can handle a huge amount of high-dimensional data and includes a feature selection algorithm [35,36,37]. Three representative DL models are deep neural networks (DNNs), convolution neural networks (CNNs), and recurrent neural networks (RNNs), and among them, RNNs, which have a loop structure and hidden states, have strength in handling sequential datasets [38]. In this study, it was difficult to apply RNN algorithms because the time series data collection interval was not constant, and the data collection frequency was different for each case. Therefore, we used DNN algorithms to extract the 12 features from time series blood laboratory data. In addition to this, RFE was used for feature selection. RFE, one of the most popular feature selection techniques, has shown excellent performances when applied with many ML and DL algorithms [19,39,40,41]. It recursively removes the weakest feature according to the ranks of the features. In this study, the DNN algorithm better predicted cancer recurrences than any of the traditional ML algorithms. When the RFE algorithm was applied to the DNN model, the ROC-AUC value was 0.883 ± 0.027 on the original training dataset, which was much higher than the ROC-AUC value of the LR (0.696 ± 0.059) with 16 features. All five selected features from the RFE algorithm were PLR-related features.
We also tried to adopt data augmentation by splitting the longitudinal blood test data for each patient. The performance of a DNN usually improves with the amount of training data. A large amount of data can also prevent overfitting problems or weak generalizability. Data augmentation is a technique that artificially creates new training data from existing training data. Various data augmentation techniques have been reported, especially for imaging data [42,43]. However, applying data augmentation on time series numerical data is difficult. The time between tumor initiation to overt presentation is not thought to take over six months. Therefore, we divided the time series data based on the 180 days before the diagnosis of recurrence for each recurrent case and created a new non-recurrent case with the data earlier than 180 days (Figure 1a). In this way, 117 non-recurrent cases were additionally generated from 206 recurrent cases. Overall, the number of data was augmented from 778 to 895 in the training dataset (Figure 1b). The cross-validation results showed comparable performances (0.889 ± 0.032) on the split (augmented) training dataset compared with those (0.883 ± 0.027) on the original training dataset. However, in external validation, the models trained with the augmented training dataset performed better than those trained with the original training dataset (Table 5). This means that data augmentation by data splitting can improve generalization performance. In addition, our model showed the ability to recognize longitudinal data samples correctly at each time point (Table 5).
In this study, we aimed to develop a DNN model to detect the recurrence after CCRT for head and neck cancers using post-treatment longitudinal laboratory data. The main limitation of this study is that the interval and amount of laboratory data are different for each case. If a larger amount of data can be obtained at regular intervals for a certain period, better ML and DL models can be developed. In addition, DNN models are often considered to be black boxes, making it challenging to understand the underlying decision-making process and resulting in a lack of interpretability. It can be difficult to explain how and why a model arrived at a particular prediction, limiting their interpretability and applications. While DNN models can learn complex representations from raw data, the learned features may not be easily interpretable or understandable by humans. This lack of transparency hinders the extraction of meaningful insights and an understanding of the underlying factors driving the models’ predictions. To address these limitations, various approaches, such as data augmentation, transfer learning, regularization methods, interpretability approaches, and adversarial training are needed to mitigate these challenges, and therefore enhance the performance and reliability of DNN models.
In this study, we extracted various features representing temporal changes in the NLR and PLR values in deep learning models and used them as input for training. Notably, applying RFE resulted in a performance of over 0.88 only with temporal changes in the PLR. With this, we interpreted the model as another follow-up tool for predicting recurrence and evaluated its value, which distinguishes our study from previous research that has mainly evaluated the predictive value of a single time point before or shortly after treatment. Furthermore, the application of RFE during the model construction process is also a significant result in terms of its technical aspect.

5. Conclusions

The DNN model with feature selection using the RFE algorithm showed the best performance among the ML models to discern a recurrence after CCRT in patients with head and neck cancers. The model performance could be enhanced with splitting and augmentation of the training dataset. The performance of the DNN-RFE model was validated with an external dataset. Systemic inflammatory markers can detect recurrence after treatment, and the predictive performance might be further improved with prospective data collection at regular intervals and periods.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers15143540/s1, Table S1: External validation dataset.

Author Contributions

Y.K.S., conceptualization, methodology, investigation, data curation, visualization, funding acquisition, writing—original draft, writing—review and editing, project administration; Z.K., formal analysis, investigation, data curation, visualization; T.Y.C., formal analysis, investigation, data curation; M.J.C., software, validation, data curation; C.-H.B., Y.-I.S., J.S., Y.-S.J., M.-J.A., Y.C.A., and D.O. resources, supervision; B.H.C., software, validation, data curation, visualization, funding acquisition, writing—review and editing; M.K.C., conceptualization, methodology, supervision, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant (NRF-2021M3E5D2A01022399), Future Medicine 20*30 Project of the Samsung Medical Center (SMO1230021 and SMC1210791), and funded by the Korean government (MSIT) (No. 2017R1C1B5018150).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Samsung Medical Center (SMC IRB file no. 2020-06-125-001).

Informed Consent Statement

Informed consent was waived due to the retrospective design.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available because there was no pre-approval by IRB.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. el-Hag, A.; Clark, R.A. Immunosuppression by activated human neutrophils. Dependence on the myeloperoxidase system. J. Immunol. 1987, 139, 2406–2413. [Google Scholar] [PubMed]
  2. Petrie, H.T.; Klassen, L.W.; Kay, H.D. Inhibition of human cytotoxic T lymphocyte activity in vitro by autologous peripheral blood granulocytes. J. Immunol. 1985, 134, 230–234. [Google Scholar] [PubMed]
  3. Sharaiha, R.Z.; Halazun, K.J.; Mirza, F.; Port, J.L.; Lee, P.C.; Neugut, A.I.; Altorki, N.K.; Abrams, J.A. Elevated preoperative neutrophil:lymphocyte ratio as a predictor of postoperative disease recurrence in esophageal cancer. Ann. Surg. Oncol. 2011, 18, 3362–3369. [Google Scholar] [PubMed] [Green Version]
  4. Chua, W.; Charles, K.A.; Baracos, V.E.; Clarke, S.J. Neutrophil/lymphocyte ratio predicts chemotherapy outcomes in patients with advanced colorectal cancer. Br. J. Cancer 2011, 104, 1288–1295. [Google Scholar]
  5. Teramukai, S.; Kitano, T.; Kishida, Y.; Kawahara, M.; Kubota, K.; Komuta, K.; Minato, K.; Mio, T.; Fujita, Y.; Yonei, T.; et al. Pretreatment neutrophil count as an independent prognostic factor in advanced non-small-cell lung cancer: An analysis of Japan Multinational Trial Organisation LC00-03. Eur. J. Cancer 2009, 45, 1950–1958. [Google Scholar] [PubMed] [Green Version]
  6. Bhatti, I.; Peacock, O.; Lloyd, G.; Larvin, M.; Hall, R.I. Preoperative hematologic markers as independent predictors of prognosis in resected pancreatic ductal adenocarcinoma: Neutrophil-lymphocyte versus platelet-lymphocyte ratio. Am. J. Surg. 2010, 200, 197–203. [Google Scholar] [CrossRef]
  7. Ding, P.R.; An, X.; Zhang, R.X.; Fang, Y.J.; Li, L.R.; Chen, G.; Wu, X.J.; Lu, Z.H.; Lin, J.Z.; Kong, L.H.; et al. Elevated preoperative neutrophil to lymphocyte ratio predicts risk of recurrence following curative resection for stage IIA colon cancer. Int. J. Colorectal Dis. 2010, 25, 1427–1433. [Google Scholar] [CrossRef]
  8. Ray-Coquard, I.; Cropet, C.; Van Glabbeke, M.; Sebban, C.; Le Cesne, A.; Judson, I.; Tredan, O.; Verweij, J.; Biron, P.; Labidi, I.; et al. Lymphopenia as a prognostic factor for overall survival in advanced carcinomas, sarcomas, and lymphomas. Cancer Res. 2009, 69, 5383–5391. [Google Scholar]
  9. Wang, D.S.; Luo, H.Y.; Qiu, M.Z.; Wang, Z.Q.; Zhang, D.S.; Wang, F.H.; Li, Y.H.; Xu, R.H. Comparison of the prognostic values of various inflammation based factors in patients with pancreatic cancer. Med. Oncol. 2012, 29, 3092–3100. [Google Scholar]
  10. Kao, S.C.; Pavlakis, N.; Harvie, R.; Vardy, J.L.; Boyer, M.J.; van Zandwijk, N.; Clarke, S.J. High blood neutrophil-to-lymphocyte ratio is an indicator of poor prognosis in malignant mesothelioma patients undergoing systemic therapy. Clin. Cancer Res. 2010, 16, 5805–5813. [Google Scholar]
  11. He, J.R.; Shen, G.P.; Ren, Z.F.; Qin, H.; Cui, C.; Zhang, Y.; Zeng, Y.X.; Jia, W.H. Pretreatment levels of peripheral neutrophils and lymphocytes as independent prognostic factors in patients with nasopharyngeal carcinoma. Head Neck 2012, 34, 1769–1776. [Google Scholar] [PubMed]
  12. Huang, S.H.; Waldron, J.N.; Milosevic, M.; Shen, X.; Ringash, J.; Su, J.; Tong, L.; Perez-Ordonez, B.; Weinreb, I.; Bayley, A.J.; et al. Prognostic value of pretreatment circulating neutrophils, monocytes, and lymphocytes in oropharyngeal cancer stratified by human papillomavirus status. Cancer 2015, 121, 545–555. [Google Scholar] [PubMed]
  13. Rassouli, A.; Saliba, J.; Castano, R.; Hier, M.; Zeitouni, A.G. Systemic inflammatory markers as independent prognosticators of head and neck squamous cell carcinoma. Head Neck 2015, 37, 103–110. [Google Scholar] [CrossRef]
  14. Rosculet, N.; Zhou, X.C.; Ha, P.; Tang, M.; Levine, M.A.; Neuner, G.; Califano, J. Neutrophil-to-lymphocyte ratio: Prognostic indicator for head and neck squamous cell carcinoma. Head Neck 2017, 39, 662–667. [Google Scholar]
  15. Rajkomar, A.; Oren, E.; Chen, K.; Dai, A.M.; Hajaj, N.; Hardt, M.; Liu, P.J.; Liu, X.; Marcus, J.; Sun, M.; et al. Scalable and accurate deep learning with electronic health records. NPJ Digit. Med. 2018, 1, 18. [Google Scholar] [CrossRef] [PubMed]
  16. Choi, E.; Bahadori, M.T.; Schuetz, A.; Stewart, W.F.; Sun, J. Doctor AI: Predicting Clinical Events via Recurrent Neural Networks. JMLR Workshop Conf. Proc. 2016, 56, 301–318. [Google Scholar] [PubMed]
  17. Kim, W.J.; Sung, J.M.; Sung, D.; Chae, M.H.; An, S.K.; Namkoong, K.; Lee, E.; Chang, H.J. Cox Proportional Hazard Regression Versus a Deep Learning Algorithm in the Prediction of Dementia: An Analysis Based on Periodic Health Examination. JMIR Med. Inform. 2019, 7, e13139. [Google Scholar]
  18. Cho, B.H.; Yu, H.; Kim, K.W.; Kim, T.H.; Kim, I.Y.; Kim, S.I. Application of irregular and unbalanced data to predict diabetic nephropathy using visualization and feature selection methods. Artif. Intell. Med. 2008, 42, 37–53. [Google Scholar]
  19. Choi, N.; Kim, Z.; Song, B.H.; Park, W.; Chung, M.J.; Cho, B.H.; Son, Y.I. Prediction of risk factors for pharyngo-cutaneous fistula after total laryngectomy using artificial intelligence. Oral Oncol. 2021, 119, 105357. [Google Scholar]
  20. Zhang, Z. Improved Adam Optimizer for Deep Neural Networks. In Proceedings of the 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS), Banff, AB, Canada, 4–6 June 2018; pp. 1–2. [Google Scholar]
  21. Balkwill, F.; Mantovani, A. Inflammation and cancer: Back to Virchow? Lancet 2001, 357, 539–545. [Google Scholar] [CrossRef]
  22. Mantovani, A.; Allavena, P.; Sica, A.; Balkwill, F. Cancer-related inflammation. Nature 2008, 454, 436–444. [Google Scholar] [PubMed]
  23. Mantovani, A.; Romero, P.; Palucka, A.K.; Marincola, F.M. Tumour immunity: Effector response to tumour and role of the microenvironment. Lancet 2008, 371, 771–783. [Google Scholar] [PubMed]
  24. Templeton, A.J.; McNamara, M.G.; Seruga, B.; Vera-Badillo, F.E.; Aneja, P.; Ocana, A.; Leibowitz-Amit, R.; Sonpavde, G.; Knox, J.J.; Tran, B.; et al. Prognostic role of neutrophil-to-lymphocyte ratio in solid tumors: A systematic review and meta-analysis. J. Natl. Cancer Inst. 2014, 106, dju124. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Zhou, X.; Du, Y.; Huang, Z.; Xu, J.; Qiu, T.; Wang, J.; Wang, T.; Zhu, W.; Liu, P. Prognostic value of PLR in various cancers: A meta-analysis. PLoS ONE 2014, 9, e101119. [Google Scholar]
  26. Takenaka, Y.; Oya, R.; Kitamiura, T.; Ashida, N.; Shimizu, K.; Takemura, K.; Yamamoto, Y.; Uno, A. Platelet count and platelet-lymphocyte ratio as prognostic markers for head and neck squamous cell carcinoma: Meta-analysis. Head Neck 2018, 40, 2714–2723. [Google Scholar] [PubMed]
  27. Kim, S.K.; Demetri, G.D. Chemotherapy and neutropenia. Hematol. Oncol. Clin. North Am. 1996, 10, 377–395. [Google Scholar]
  28. Meyer, K.K. Radiation-induced lymphocyte-immune deficiency. A factor in the increased visceral metastases and decreased hormonal responsiveness of breast cancer. Arch. Surg. 1970, 101, 114–121. [Google Scholar]
  29. Chowdhary, M.; Switchenko, J.M.; Press, R.H.; Jhaveri, J.; Buchwald, Z.S.; Blumenfeld, P.A.; Marwaha, G.; Diaz, A.; Wang, D.; Abrams, R.A.; et al. Post-treatment neutrophil-to-lymphocyte ratio predicts for overall survival in brain metastases treated with stereotactic radiosurgery. J. Neurooncol. 2018, 139, 689–697. [Google Scholar]
  30. Suh, K.J.; Kim, S.H.; Kim, Y.J.; Kim, M.; Keam, B.; Kim, T.M.; Kim, D.W.; Heo, D.S.; Lee, J.S. Post-treatment neutrophil-to-lymphocyte ratio at week 6 is prognostic in patients with advanced non-small cell lung cancers treated with anti-PD-1 antibody. Cancer Immunol. Immunother. 2018, 67, 459–470. [Google Scholar]
  31. Zhuang, Y.; Yuan, B.Y.; Hu, Y.; Chen, G.W.; Zhang, L.; Zhao, X.M.; Chen, Y.X.; Zeng, Z.C. Pre/Post-Treatment Dynamic of Inflammatory Markers Has Prognostic Value in Patients with Small Hepatocellular Carcinoma Managed by Stereotactic Body Radiation Therapy. Cancer Manag. Res. 2019, 11, 10929–10937. [Google Scholar]
  32. Zhou, C.; Hu, J.; Wang, Y.; Ji, M.H.; Tong, J.; Yang, J.J.; Xia, H. A machine learning-based predictor for the identification of the recurrence of patients with gastric cancer after operation. Sci. Rep. 2021, 11, 1571. [Google Scholar] [PubMed]
  33. Rahman, S.A.; Walker, R.C.; Lloyd, M.A.; Grace, B.L.; van Boxel, G.I.; Kingma, B.F.; Ruurda, J.P.; van Hillegersberg, R.; Harris, S.; Parsons, S.; et al. Machine learning to predict early recurrence after oesophageal cancer surgery. Br. J. Surg. 2020, 107, 1042–1052. [Google Scholar] [PubMed] [Green Version]
  34. Xu, Y.; Ju, L.; Tong, J.; Zhou, C.M.; Yang, J.J. Machine Learning Algorithms for Predicting the Recurrence of Stage IV Colorectal Cancer After Tumor Resection. Sci. Rep. 2020, 10, 2519. [Google Scholar] [PubMed] [Green Version]
  35. Wainberg, M.; Merico, D.; Delong, A.; Frey, B.J. Deep learning in biomedicine. Nat. Biotechnol. 2018, 36, 829–838. [Google Scholar]
  36. Park, Y.; Kellis, M. Deep learning for regulatory genomics. Nat. Biotechnol. 2015, 33, 825–826. [Google Scholar]
  37. Kim, D.W.; Lee, S.; Kwon, S.; Nam, W.; Cha, I.H.; Kim, H.J. Deep learning-based survival prediction of oral cancer patients. Sci. Rep. 2019, 9, 6994. [Google Scholar]
  38. Chatzikonstantinou, C.; Konstantinidis, D.; Dimitropoulos, K.; Daras, P. Recurrent neural network pruning using dynamical systems and iterative fine-tuning. Neural Netw. 2021, 143, 475–488. [Google Scholar]
  39. Senan, E.M.; Al-Adhaileh, M.H.; Alsaade, F.W.; Aldhyani, T.H.H.; Alqarni, A.A.; Alsharif, N.; Uddin, M.I.; Alahmadi, A.H.; Jadhav, M.E.; Alzahrani, M.Y. Diagnosis of Chronic Kidney Disease Using Effective Classification Algorithms and Recursive Feature Elimination Techniques. J. Healthc. Eng. 2021, 2021, 1004767. [Google Scholar]
  40. Chen, Q.; Meng, Z.; Su, R. WERFE: A Gene Selection Algorithm Based on Recursive Feature Elimination and Ensemble Strategy. Front. Bioeng. Biotechnol. 2020, 8, 496. [Google Scholar]
  41. Karami, G.; Giuseppe Orlando, M.; Delli Pizzi, A.; Caulo, M.; Del Gratta, C. Predicting Overall Survival Time in Glioblastoma Patients Using Gradient Boosting Machines Algorithm and Recursive Feature Elimination Technique. Cancers 2021, 13, 4976. [Google Scholar]
  42. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [PubMed]
  43. He, C.; Liu, J.; Zhu, Y.; Du, W. Data Augmentation for Deep Neural Networks Model in EEG Classification Task: A Review. Front. Hum. Neurosci. 2021, 15, 765525. [Google Scholar] [PubMed]
Figure 1. Sample split for recurrent cases. (a) For each recurrent case, the temporal blood laboratory values were split into two groups based on 180 days before the diagnosis of recurrence. (b) With data splitting for the recurrent cases, a modified dataset, named split training dataset, was built (N = 895).
Figure 1. Sample split for recurrent cases. (a) For each recurrent case, the temporal blood laboratory values were split into two groups based on 180 days before the diagnosis of recurrence. (b) With data splitting for the recurrent cases, a modified dataset, named split training dataset, was built (N = 895).
Cancers 15 03540 g001aCancers 15 03540 g001b
Figure 2. The receiver operating characteristic (ROC) curve analysis showing the performances of each machine learning model. (a) analysis results from the original dataset. (b) analysis results from the split dataset.
Figure 2. The receiver operating characteristic (ROC) curve analysis showing the performances of each machine learning model. (a) analysis results from the original dataset. (b) analysis results from the split dataset.
Cancers 15 03540 g002aCancers 15 03540 g002b
Figure 3. Performances of AI models from the original and split datasets: (a) The RFE algorithm is applied to the DNN model in the original (non-split) training dataset, and the top three features (PLR mean, PLR Tx, and PLR max–min) were selected. With these three features, the ROC-AUC and the PR-AUC values of the DNN-RFE model were 0.884 ± 0.032 and 0.788 ± 0.032, respectively, in the original training dataset; (b) the DNN-RFE algorithm is applied in a split training dataset, and the PLR R, PLR R-Tx, PLR mean, PLR Tx, and PLR max–min were selected. With the top five features, the ROC-AUC and the PR-AUC values of the DNN-RFE model were 0.889 ± 0.032 and 0.771 ± 0.044, respectively, in the split training dataset. When these top five features were used in the original training dataset, the ROC-AUC was 0.883 ± 0.027 (Table 3(a)).
Figure 3. Performances of AI models from the original and split datasets: (a) The RFE algorithm is applied to the DNN model in the original (non-split) training dataset, and the top three features (PLR mean, PLR Tx, and PLR max–min) were selected. With these three features, the ROC-AUC and the PR-AUC values of the DNN-RFE model were 0.884 ± 0.032 and 0.788 ± 0.032, respectively, in the original training dataset; (b) the DNN-RFE algorithm is applied in a split training dataset, and the PLR R, PLR R-Tx, PLR mean, PLR Tx, and PLR max–min were selected. With the top five features, the ROC-AUC and the PR-AUC values of the DNN-RFE model were 0.889 ± 0.032 and 0.771 ± 0.044, respectively, in the split training dataset. When these top five features were used in the original training dataset, the ROC-AUC was 0.883 ± 0.027 (Table 3(a)).
Cancers 15 03540 g003
Table 1. The definition of 12 blood laboratory meta-features.
Table 1. The definition of 12 blood laboratory meta-features.
FeaturesDefinition
NLR mean, PLR meanMean NLR (PLR)
NLR max–min, PLR max–minMaximum NLR (PLR) minus minimum NLR (PLR)
NLR Tx, PLR TxNLR (PLR) after treatment *
NLR R, PLR RNLR (PLR) at the time of recurrence or at the end of the period **
NLR R-mean, PLR R-meanNLR R (PLR R) minus NLR mean (PLR mean)
NLR R-Tx, PLR R-TxNLR R (PLR R) minus NLR Tx (PLR Tx)
* NLR Tx (PLR Tx): NLR (PLR) after treatment (values from 14 days to 90 days after the end of treatment). ** NLR R (PLR R): NLR (PLR) at the time of recurrence for recurrent cases and at the end of the period for non-recurrent cases.
Table 2. Study patients (N = 778).
Table 2. Study patients (N = 778).
CharacteristicsValues
Age (year) Mean ± SD60.4 ± 13.0
Gender (male/female) 625/153
Smoking 481 (61.8%)
Stage *I11 (1.4%)
II68 (8.8%)
III209 (26.9%)
IV489 (62.9%)
Tumor sitesOral cavity 45 (5.8%)
Larynx54 (6.9%)
Oropharynx228 (29.3%)
Hypopharynx74 (9.5%)
Nasopharynx296 (38.0%)
PNS/nasal cavity42 (5.4%)
Others39 (5.0%)
PathologySCC503 (64.7%)
UDC243 (31.2%)
Others32 (4.1%)
Follow-up period (months)Median (interquartile range)41.8 (17.4~71.6)
Recurrences 206 (26.5%)
RFS (months)Median (SE, 95% CI)189.8 (33.5, 124.2~255.3)
5-year RFS rate 66.5%
* Staging according to the AJCC 7th edition. The value for one case was missing. SD, standard deviation; SCC, squamous cell carcinoma; UDC, undifferentiated carcinoma; RFS, relapse-free survival; SE, standard error.
Table 3. The performance of the trained models.
Table 3. The performance of the trained models.
a. from Original Dataset
ModelTrainingOriginal Dataset
Features (N)ROC-AUCPR-AUCAccuracyPrecisionSensitivitySpecificity
Logistic regression160.696 ± 0.0590.454 ± 0.0790.770 ± 0.0390.606 ± 0.1050.885 ± 0.0430.461 ±0.108
Random forest160.812 ± 0.0490.622 ± 0.0810.855 ± 0.0270.758 ± 0.0580.917 ± 0.0260.687 ± 0.081
Gradient boosting160.600 ± 0.0520.346 ± 0.0550.825 ± 0.0430.744 ± 0.1160.926 ± 0.0380.554 ± 0.065
DNN160.828 ± 0.0320.663 ± 0.0690.686 ± 0.1020.471 ± 0.1120.675 ± 0.0700.715 ± 0.070
DNN-RFE50.883 ± 0.0270.778 ± 0.0420.801 ± 0.0300.600 ± 0.0410.800 ± 0.0530.801 ± 0.053
b. from Split Dataset
ModelTrainingSplit Dataset
Features (N)ROC-AUCPR-AUCAccuracyPrecisionSensitivitySpecificity
Logistic regression160.700 ± 0.0530.456 ± 0.0720.759 ± 0.0270.482 ± 0.2000.951 ± 0.0290.139 ± 0.070
Random forest160.751 ± 0.0530.539 ± 0.0840.845 ± 0.0280.737 ± 0.0840.940 ± 0.0230.536 ± 0.082
Gradient boosting160.735 ± 0.0570.516 ± 0.0850.822 ± 0.0460.731 ± 0.1140.922 ± 0.0380.554 ± 0.079
DNN160.837 ± 0.0380.666 ± 0.0450.786 ± 0.0260.537 ± 0.0440.808 ± 0.0290.715 ± 0.029
DNN-RFE50.889 ± 0.0320.771 ± 0.0440.812 ± 0.0140.578 ± 0.0220.827 ± 0.0940.763 ± 0.094
DNN, deep neural network; DNN-RFE, deep neural network-recursive feature elimination; ROC-AUC, receiver operating curve-area under the curve; PR-AUC, precision recall-area under the curve.
Table 4. Feature selection of the DNN model using the RFE algorithm.
Table 4. Feature selection of the DNN model using the RFE algorithm.
No. of FeaturesEliminated FeatureOriginal Training DatasetSplit Training Dataset
ROC-AUCPR-AUCROC-AUCPR-AUC
16Sex0.828 0.663 0.8370.666
15Age0.836 0.676 0.8170.659
14Smoking0.864 0.725 0.8540.732
13Stage0.867 0.750 0.8620.693
12NLR max–min0.867 0.731 0.8750.756
11NLR R0.861 0.731 0.8870.752
10NLR Tx0.849 0.712 0.8690.738
9NLR mean0.858 0.732 0.8660.730
8NLR R-mean0.865 0.753 0.8720.739
7NLR R-Tx0.872 0.766 0.8860.749
6PLR R-mean0.881 0.763 0.8890.764
5PLR R0.883 0.778 0.8890.771
4PLR R-Tx0.883 0.764 0.8850.765
3PLR mean0.884 0.788 0.8800.758
2PLR Tx0.785 0.582 0.7520.444
1PLR max–min0.724 0.550 0.6970.418
DNN, deep neural network; RFE, recursive feature elimination; ROC-AUC, receiver operating curve-area under the curve; PR-AUC, precision recall-area under the curve.
Table 5. External validation and cross-validation performance.
Table 5. External validation and cross-validation performance.
Training DatasetSplit TrainingOriginal TrainingSplit TrainingOriginal Training
Validation DatasetSplit ValidationSplit ValidationOriginal ValidationOriginal Validation
ModelFeaturesROC-AUCPR-AUCROC-AUCPR-AUCROC-AUCPR-AUCROC-AUCPR-AUC
LR160.5710.4040.5820.4060.5560.4810.5720.494
RF160.5310.3520.5380.3560.5310.4490.7480.702
GB160.7170.5500.5950.4120.6720.5960.5850.496
DNN160.8320.7370.5580.3950.7660.7660.6430.560
DNN-RFE50.8450.7850.7080.5450.7840.7230.7100.670
LR, logistic regression; RF, random forest; GB, gradient boosting; DNN, deep neural network; DNN-RFE, deep neural network-recursive feature elimination.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

So, Y.K.; Kim, Z.; Cheong, T.Y.; Chung, M.J.; Baek, C.-H.; Son, Y.-I.; Seok, J.; Jung, Y.-S.; Ahn, M.-J.; Ahn, Y.C.; et al. Detection of Cancer Recurrence Using Systemic Inflammatory Markers and Machine Learning after Concurrent Chemoradiotherapy for Head and Neck Cancers. Cancers 2023, 15, 3540. https://doi.org/10.3390/cancers15143540

AMA Style

So YK, Kim Z, Cheong TY, Chung MJ, Baek C-H, Son Y-I, Seok J, Jung Y-S, Ahn M-J, Ahn YC, et al. Detection of Cancer Recurrence Using Systemic Inflammatory Markers and Machine Learning after Concurrent Chemoradiotherapy for Head and Neck Cancers. Cancers. 2023; 15(14):3540. https://doi.org/10.3390/cancers15143540

Chicago/Turabian Style

So, Yoon Kyoung, Zero Kim, Taek Yoon Cheong, Myung Jin Chung, Chung-Hwan Baek, Young-Ik Son, Jungirl Seok, Yuh-Seog Jung, Myung-Ju Ahn, Yong Chan Ahn, and et al. 2023. "Detection of Cancer Recurrence Using Systemic Inflammatory Markers and Machine Learning after Concurrent Chemoradiotherapy for Head and Neck Cancers" Cancers 15, no. 14: 3540. https://doi.org/10.3390/cancers15143540

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop