Prognostic Models for Oral Squamous Cell Carcinoma: Comparison
Please note this is a comparison between Version 1 by Giuseppe Troiano and Version 2 by Lindsay Dong.

An accurate prediction of cancer survival is very important for counseling, treatment planning, follow-up, and postoperative risk assessment in patients with Oral Squamous Cell Carcinoma (OSCC). There has been an increased interest in the development of clinical prognostic models and nomograms which are their graphic representation.

  • oral squamous cell carcinoma
  • prognosis
  • prognostic model
  • personalized medicine

1. Background

Head and Neck Cancer (HNC) is the sixth most common type of cancer across the world with nearly 550,000 new cases per year. Most of HNCs are diagnosed as Oral Squamous Cell Carcinomas (OSCC) and oral cancer ranks eighth among the most common causes of cancer-related deaths worldwide [1][2][1,2]. Both pharmacological and surgical protocols for OSCCs diagnosed in early stages are less aggressive and characterized by better outcomes, whilst in advanced stages, very high patients’ morbidity and poor clinical outcomes are expected [3]. Despite the increased knowledge and the encouraging scientific findings of the past 20 years on such diseases, the overall 5-year survival rate for OSCC is still below 50% [4].
Nowadays, the Tumor-Node-Metastasis (TNM) staging system is employed worldwide to predict tumor prognosis and to guide physicians towards the correct treatment choice, however, survival outcomes in patients classified within the same TNM stage class could be dramatically different, with discrepancies in therapy response and tumor management [5].
One of the main limitations of OSCC-related TNM system is its main focus on the anatomical extension of the disease. However, within each staging group, the prognosis can be modified by tumor-related factors, such as genetics, patient age, sex, race or comorbidities. For this reason, the need for a more “personalized” approach to the oncologic patient was underlined in the recent eighth edition of the American Joint Committee On Cancer (AJCC) staging system [6]. It is, therefore, necessary to investigate further prognostic factors to construct prognostic models to carry out a personalized prognosis evaluation [7][8][7,8].
Recently, there has been an increased interest in the development of clinical prognostic models and, in particular, in nomograms which are their graphic representation [9]. These are a set of mathematical algorithms that can be used to predict patient outcomes by incorporating multiple variables. Clinic-pathological and genetic variables are mainly incorporated in OSCC prognostic models, showing interesting evidence of their role in patients’ prognosis [10][11][10,11]. Purpose of these models is to estimate the probability or individual risk that a given condition, such as recurrence or death, will occur in a specific time by combining information from multiple prognostic factors of an individual [12].
Due to the recent interest in these new prognostic tools, and their potential important role in clinical practice, some guidelines have been defined for explanation and elaboration of clinically useful and correctly elaborated prognostic model. These Guidelines are reported in the Prognosis Research Strategy (PROGRESS) 3 and the Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD) [7][13][7,13]. In 2016 the AJCC developed the acceptance criteria for inclusion of risk models for individualized prognosis in the practice of precision medicine in the systematic reviews [14]. In the same year, Debray et al. developed a guide for systematic reviews and meta-analyzes of the performance of prognostic models [15]. Additionally, the Prediction Model Risk of Bias Assessment Tool (PROBAST) was also developed to assess the risk of bias and the applicability of diagnostic and prognostic prediction model studies [16].

2. Prognostic Models for Oral Squamous Cell Carcinoma

Methodological characteristics of prognostic models developed are summarized on Table 13.

Table 13.
Methodological characteristics of prognostic models developed.
Authors and Year Internal Validation Modelling Method Handling of Missing Data Model Discrimination Model Calibration Model Presentation Handling of Continuous Predictors Non-Linearity Internal Validation C-Index External Validation C-Index
Bobdey 2016

[17]
Bobdey 2016

[20]
1000-time bootstrapping Multivariable Cox proportional hazards regression models and stepdown reduction method n/a C-statistic n/a Nomogram Mixed: Continuous; Categorical/dichotomous none 0.7263 none
Li

2017

[18]
Li

2017

[21]
1000-time bootstrapping Multivariable Cox proportional hazards regression models n/a C-statistic Calibration plot Nomogram Categorical/dichotomous n/a 0.709 0.691
Montero

2014

[19]
Montero

2014

[22]
1000-time bootstrapping Multivariable Cox proportional hazards regression models and stepdown reduction method Imputation C-statistic Calibration plot Nomogram Categorical/dichotomous Cubic splines 0.67 none
Sun

2019

[20]
Sun

2019

[23]
Combination of methods: 500-time bootstrapping; 5-fold cross-validation Multivariable Cox proportional hazards regression models n/a C-statistic Calibration plot Nomogram Mixed: Continuous; Categorical/dichotomous none 0.705 0.664
Bobdey

2017

[21]
Bobdey

2017

[24]
1000-time bootstrapping Multivariable Cox proportional hazards regression models and stepdown reduction method n/a C-statistic n/a Nomogram Categorical/dichotomous n/a 0.7266 0.740
Chang

2018

[22]
Chang

2018

[25]
1000-time bootstrapping Multivariable Cox proportional hazards regression models n/a AUC Calibration plot Nomogram Categorical/dichotomous Cubic splines 0.78 0.71

An accurate prediction of cancer survival is very important for counseling, treatment planning, follow-up and postoperative risk assessment in patients with OSCC [23][27]. Although the use of prognosis models is still relatively new for OSCC, these models are already widely used for other human diseases [24][25][26][27][28,29,30,31]. It is now well known that cancer-related outcomes are influenced by several factors that are not included in the TNM system. The vast majority of these factors has not been incorporated into the staging system because they may not predict outcome “independently” in multivariate prognosis models, however many of them may work in tandem and have varying degrees of influence on each other [28][29][32,33].

Six studies included correctly developed models according to the TRIPOD, all the included studies carried out internal validation of the model and four models were also externally validated [18][19][20][21][22][30][21,22,23,24,25,26]. The majority of models assessed OS in patients with squamous cell carcinoma of the tongue [19][21][30][22,24,26], two assessed all possible sites of tumor onset [18][20][21,23], and one model only assessed the buccal mucosa cancer [22][25]. All models rated OS at five years, except for Bobdey et al [22][25]. who only rated it at three years; furthermore, Li et al. and Sun et al., also evaluated OS at eight and three years respectively [18][20][21,23]. Among the clinical factors, those most included in the models are age, race, martial state, comorbidities and smoking; while among the histopathological ones the most investigated were T stage, N stage and M stage.

It is well known that the performance of a prognostic model is overestimated when it is just assessed in the patient sample that was used to build the model [31][34]. Internal validation provides a better estimate of model performance in new patients when done by adjusting overfitting, that is the difference between the accuracy of the apparent prediction and the accuracy of the prediction measured on an independent test set. Resampling techniques are a set of methods to provide an assessment of accuracy for the developed prognostic prediction models [32][35]. As an exception, Sun et al. [20][23] used a combined bootstrapping and cross-validation method, although all other studies used 1000-time bootstrapping as a resampling technique. Nevertheless, an evaluation of a model’s performance by using bootstrapping or cross-validation is not enough to overcome overfitting, such type of studies should also apply shrinkage, which is a method used adjust the regression coefficients [33][34][36,37].

Calibration reflects the agreement between the model’s predictions and the observed outcomes. It is preferably reported graphically, usually with a calibration plot [35][39]. Another key aspect of the characterization of a prognostic model is discrimination, that is, the ability of a forecasting model to differentiate between those who experience the outcome event or not [13]. The most used measure for discrimination is the Concordance Index (C-index), which reflects the probability that for any pair of individuals randomly, one with and one without the outcome, the model assigns a higher probability to the individual with the outcome [36][40]. For survival models, many c-indices have been proposed, so it is important to underline that, from our results, the most commonly used is the discrimination model proposed by Harrell [37][41]. In any case, discrimination can vary in a range from 0 to 1 and is considered good when higher than 0.5, considering that all the studies included in this systematic review presented a C-index at least higher than 0.6, all of them showed a good prognostic accuracy [38][42]. In addition, improvements in study design and analysis are crucial to allow evidence of more reliable prognostic factors that can be incorporated into new prognostic models, or to update existing models, to improve discrimination [39][43]. Another important finding was the almost total lack of handling of the missing data, except for Montero et al. [19][22] who carried out the multivariate imputations by chained equations (MICE) [40][44] before conducting multivariable regression statistical analysis [20][23]. The absence of a mention of the missing data leads to a so-called “full case analysis”. Including only participants with complete data, as well as being inefficient as it reduces the sample, can also lead to biased results due to a subsample [12].

External validation is preferable to internal validation for testing the transportability of a model since it is impossible for the population, or distribution of predictors, in an independent population to be the same as in the model development population [41][45]. Secondly, to improve the generalizability of a model, it should ideally be validated in different contexts with different population [42][46]. Furthermore, in the literature, there are currently no external validation by independent researchers of prognostic models for OS in patients with OSCC. A reliable model should be tested by independent researchers in different contexts to ensure the generalizability of prognostic models [15].

Most of the prognostic models in the literature describe the development of the model, a small number report external validation studies and currently, there are no studies considering clinical impact or utility [7]. Identifying accurate prognostic models and performing impact studies to investigate their influence on decision making, patient outcomes and costs is a fundamental component of stratified medicine because it contributes evidence at multiple stages in translation [43][47].
Multivariable Cox proportional hazards regression models were used to developing the models, as indicated for survival data [44][48]. All included prognostic models used nomogram as model presentation, yet none of the prognostic models reported the original mathematical regression formula. This turns out to be highly limiting, firstly because this presentation format is not a simplification of a developed model, but rather a graphical presentation of the original mathematical regression formula, and secondly, because recalibration, and updating of the original formula is necessary to perform validation [45][49]. Furthermore, it would be advisable to provide readers with the appropriate tools for the interpretation and application of the nomogram [26][30].

3. Conclusions

The following recommendations could be reported: (i) model development studies should weight for overfitting by carrying out internal validation (by resampling techniques such as bootstrapping) and using shrinkage techniques, (ii) model calibration and discrimination should always be examined, (iii) imputation techniques for missing data handling should always be applied, (iv) non-linearity of continuous predictors should be examined, (v) the complete equation of the prognostic model should always be reported to allow external validation and updating by independent research groups; (vi) prospective studies should be performed to reduce the risk of bias (vii) external validation in a new context and impact assessment on health outcomes and cost effectiveness of care should be carried out.