Page 522 - Decision Making Applications in Modern Power Systems
P. 522

A reliable decision-making algorithm Chapter | 18  481



                   Input       Bootstrapping     Bagging       Aggregation


                                  Sample (1)    Decision tree-1


                                                                 Ensembler
                   Dataset        Sample (2)    Decision tree-2
                                                                 (Avg/Vote)
                                 Sample (N)     Decision tree-N


             FIGURE 18.5 Architecture of a BDT model. BDT, Bagged decision tree.

             such tree-based ensemble technique is BDT that builds, by randomization,
             numerous DTs and then aggregates their predictions as exemplified in
             Fig. 18.5. From an ensemble of trees, one can derive an importance score for
             each variable of the problem that assesses its relevance for predicting the out-
             put. Every tree in the ensemble is grown on an independently drawn bootstrap
             replica of input data [29]. Observations not included in this replica are “out-
             of-bag” for this tree [30]. The bagging is nothing but a bootstrap aggregation
             of a set of DTs. Individual DTs tend to over-fit. BDT (Bootstrap aggregated)
             combines the results of many DTs, thereby overwhelming the effects of
             over-fitting and improves generalization. Tree Bagger grows the DTs in the
             ensemble using bootstrap samples of the data. The estimated out-of-bag error
             variances can be reduced by setting a more balanced misclassification cost
             matrix or a less skewed prior probability vector. BDT generates in-bag
             samples by oversampling classes with large misclassification costs and under-
             sampling classes with small misclassification costs. Consequently, out-of-bag
             samples have fewer observations from classes with large misclassification
             costs and more observations from classes with small misclassification costs.
             To train a BDT ensemble classifier using a small data set, a highly skewed
             cost matrix is considered, and then the number of out-of-bag observations per
             class might be very low. In both bagging and random forest, many individual
             DTs are built on a bootstrapped version of the original dataset and are ensem-
             ble together. The difference is that, in a random forest, a random subset of
             variables are considered during node split while in bagging of DTs, all the
             variables are considered in a node split [31].


             18.4.3 Wavelet packet energy and bagged decision tree based
             decision-making algorithm

             In order to train and test the proposed WPE_BDT based decision-making
             algorithm for detection and classification of shunt faults during PS condition,
   517   518   519   520   521   522   523   524   525   526   527