Archives

  • 2019-07
  • 2019-08
  • 2019-09
  • 2019-10
  • 2019-11
  • 2020-03
  • 2020-07
  • 2020-08
  • 2021-03
  • Continuous br Table br UCI datasets used for evaluating perf

    2019-10-02

    Continuous 5.19 31.73 1102.50
    Table 4
    UCI datasets used for evaluating performances of different online learning models.
    Dataset
    Data Optimization Training & Number of
    Size Set
    Testing Set Batches
    WDBC
    Table 5
    Comparison of 5 online learning models on 3 UCI datasets.
    Performance
    Ionosphere
    WBCD Spambase
    Metric Model Average SD Average SD Average SD
    Fig. 3. Performances of online models on SEER BC dataset.
    The GAOGB model shows a testing accuracy of 75.0% in aver-age and 75.2% in median. The AUC with GAOGB USP7/USP47 inhibitor 75.1% in average and 75.2% in median. As demonstrated in Fig. 3, GAOGB outper-forms other online models significantly. In addition, GAOGB shows the lowest variations with Accuracy and AUC. Even though OSELM was the second best model in terms of Accuracy and AUC with UCI datasets, its advantage faded with SEER data, noted by a 71.9% testing accuracy. Instead, OLRGB model, with 73.8% testing accu-racy, performs second best among all models with SEER data. Still, GAOGB is more accurate than the OLRGB from a two sample test (p-value < 0.01).
    To understand the advantages of GAOGB compared to conven-tional o ine learning algorithms, Fig. 4 illustrates a comparison
    Table 6
    pretation of results, the dataset is balanced. The survival cases are under sampled to gain 5,991 cases equally for survival and non-survival cases, amounting to 11,982 observations for compu-tation. The balanced SEER dataset is split to obtain 10% (1200 sam-ples) as the optimization set, 72% (8627 samples) as the training set and 18% (2155 samples) as the testing set. The training set is further split into 10 batches to formulate an online learning environment.
    4.5. Evaluation on SEER dataset
    The GAOGB model is comprehensively evaluated by comparing to both online and o ine models. Fig. 3 (online models) and Fig. 4 (o ine models) represent the evaluation results in 5 replications on SEER BC dataset. The averaged results for all replications of on-line models are given in Table 7. 
    Range of each attributes for preprocessed SEER Breast Cancer dataset.
    Categorical variable Number of unique values
    Race
    Marital status
    Primary site code
    Histology
    Behavior
    Grade
    Extension of disease
    Lymph node involvement
    Stage of cancer
    Site specific surgery code
    Continuous variable Mean SD Range
    Table 7
    Experimental results on the SEER BC prognosis dataset with different models.
    Specificity (%) Sensitivity (%) Retraining Time (Sec.)
    Average SD
    Average SD Average SD Average SD Average SD
    Fig. 4. Comparison between GAOGB and batch learning algorithms as the number of training batches increase.
    between GAOGB and o ine algorithms, including the SVM using the Radial Basis Kernel Function, the MLP and the AdaBoost meth-ods. By dividing the training data into 5 batches, Fig. 4 represents the changes in accuracy and retraining time with the number of batch data used for training. The GAOGB model shows insignifi-cant difference to the AdaBoost model in terms of accuracy. With a 75.9% accuracy after training 5 batches, GAOGB shows a trade-off of 0.7% in accuracy compared to the final 76.6% accuracy of the SVM and MLP models. However, the dominating advantages of GAOGB in retraining e ciency is significant in Fig. 4(b). As the number of batches increase, GAOGB maintains a retraining time around 30 s, while the retraining times of the o ine models keep increasing. At Iteration 5, the required retraining time of GAOGB is 24 times lower than the SVM model, 4 times lower than the MLP model, and 3 times lower than the AdaBoost model. The weighted area under the ROC curve ensemble (WAUCE) model proposed by Wang et al. (2017) provides another source of comparison. As a batch learning algorithm, with the same data preprocessing meth-ods, the WAUCE model achieved 76.4% accuracy (Wang et al., 2017). With a 1% average trade-off in accuracy, GAOGB significantly reduced the retraining redundancy of batch learning models.