Prévia do material em texto
J. Cent. South Univ. (2021) 28: 527−542
DOI: https://doi.org/10.1007/s11771-021-4619-8
Rockburst prediction in hard rock mines developing bagging and
boosting tree-based ensemble techniques
WANG Shi-ming(王世鸣)1, ZHOU Jian(周健)2, LI Chuan-qi(李传奇)2,
Danial Jahed ARMAGHANI3, LI Xi-bing(李夕兵)2, Hani S. MITRI4
1. School of Civil Engineering, Hunan University of Science and Technology, Xiangtan 411201, China;
2. School of Resources and Safety Engineering, Central South University, Changsha 410083, China;
3. Department of Civil Engineering, Faculty of Engineering, University of Malaya,
50603 Kuala Lumpur, Malaysia;
4. Department of Mining and Materials Engineering, McGill University, Montreal, Canada
© Central South University Press and Springer-Verlag GmbH Germany, part of Springer Nature 2021
Abstract: Rockburst prediction is of vital significance to the design and construction of underground hard rock mines.
A rockburst database consisting of 102 case histories, i.e., 1998−2011 period data from 14 hard rock mines was
examined for rockburst prediction in burst-prone mines by three tree-based ensemble methods. The dataset was
examined with six widely accepted indices which are: the maximum tangential stress around the excavation boundary
(MTS), uniaxial compressive strength (UCS) and uniaxial tensile strength (UTS) of the intact rock, stress concentration
factor (SCF), rock brittleness index (BI), and strain energy storage index (EEI). Two boosting (AdaBoost.M1, SAMME)
and bagging algorithms with classification trees as baseline classifier on ability to learn rockburst were evaluated. The
available dataset was randomly divided into training set (2/3 of whole datasets) and testing set (the remaining datasets).
Repeated 10-fold cross validation (CV) was applied as the validation method for tuning the hyper-parameters. The
margin analysis and the variable relative importance were employed to analyze some characteristics of the ensembles.
According to 10-fold CV, the accuracy analysis of rockburst dataset demonstrated that the best prediction method for
the potential of rockburst is bagging when compared to AdaBoost.M1, SAMME algorithms and empirical criteria
methods.
Key words: rockburst; hard rock; prediction; bagging; boosting; ensemble learning
Cite this article as: WANG Shi-ming, ZHOU Jian, LI Chuan-qi, Danial Jahed ARMAGHANI, LI Xi-bing, Hani S.
MITRI. Rockburst prediction in hard rock mines developing bagging and boosting tree-based ensemble techniques [J].
Journal of Central South University, 2021, 28(2): 527−542. DOI: https://doi.org/10.1007/s11771-021-4619-8.
1 Introduction
Rockburst as a common geological and
dynamic hazard commonly occurs in the
underground hard rock mines [1−3]. The
occurrence of rockburst is caused by the release of
accumulated energy in the rock in a violent way [4].
Rockbursts occurred suddenly and intensely which
usually results in considerable damage to
equipment/infrastructure and may even bring about
injuries and fatalities [5−7]. Nowadays, with the
scarcity of minerals in the shallower formations,
mining must move farther from the surface and
Foundation item: Projects(41807259, 51604109) supported by the National Natural Science Foundation of China; Project(2020CX040)
supported by the Innovation-Driven Project of Central South University, China; Project(2018JJ3693) supported by the
Natural Science Foundation of Hunan Province, China
Received date: 2020-07-13; Accepted date: 2020-09-04
Corresponding author: ZHOU Jian, PhD, Associate Professor; Tel: +86-18175162802; E-mail: csujzhou@hotmail.com; ORCID: https://
orcid.org/0000-0003-4769-4487
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
528
more widely, which could make the problem of
rockburst worse [2, 8]. It is imperative to know
whether rock excavations would burst or not. Thus,
the rockburst prediction with high accuracy is
crucial for reducing the risk of rockburst hazards
and improving the level of mining safety in
preliminary design. Although it is not easy to
precisely forecast the rockburst during activies of
mining, in the past several decades, extensive
rockburst studies have been implemented in China,
South Africa, Australia, Canada, and many other
countries [2, 3, 9−13]. Many researchers have
published a great deal of valuable results by ways
of the electromagnetic radiation method,
microseismicity monitoring, empirical criteria
classification, in situ testing methods, as well as
probabilistic methods on predicting rockbursts
[1, 2, 5, 10, 13−15]. Moreover, various types of
empirical methods and preliminary & qualitative
judgment prediction methods investigated the
mechanical characters of rockbursts through
combining local monitoring data and laboratory
tests, and are often applied in practice engineering
design. These collective efforts have greatly
improved the understanding of rockbursts. As
pointed out by ZHOU et al [2], however, universal
and practical rockburst criteria are rather difficult
endorsed in hard rock mines.
Besides the abovementioned works, many
statistical machine learning (ML)-based approaches
for rockburst prediction have been investigated
during recent decades. Most cited data-driven
researches are reported in Figure 1 [2, 5, 13, 16−23].
Conventional discriminant analysis (DA)-based
techniques such as Mahalanobis distance DA [18],
Fisher DA [19], quadratic and partial least squares
DA [13] are among the most commonly-used for
rockburst classification based on real case histories.
However, the disadvantage of the DA classifier is
that it is suitable for a class of data with a unimodal
Gaussian distribution and therefore, can only be
successfully applied to such scenarios. Moreover,
many supervised learning (SL) techniques including
adaptive neural fuzzy inference system (ANFIS),
artificial neural network (ANN), decision tree,
random forest (RF), naive Bayes (NB), K-nearest
neighbor (KNN), support vector machine (SVM),
and gradient boosting machine (GBM) have been
applied in this field. These models have been
confirmed the ability to operate on sets of input and
Figure 1 Most cited applications of data-driven approaches for rockburst prediction
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
529
output data from rockburst case histories.
Compared with traditional prediction models, the
greatest strength of SL modeling techniques is their
high efficiency to capture the nonlinear
relationships between features of the dataset instead
of assuming a preconceived interaction between
inputs and output (s). Although these methods are
able to achieve satisfactory results, they have
several identified shortcomings. For example, the
potential shortcomings of ANN include slow
learning rate and falls into local minima [24]. The
modeling with ANFIS model and fining the best
membership functions and rules of inputs is
time-consuming [21, 25−27]. The SVM classifier is
limited by the increase in the number of training
vectors which lead to computation and storage
requirements increased rapidly [28]. In large
processingtime, kNN algorithm has some
limitations to classify a new unknown observation
and difficult to improve the classification accuracy
when deal with the multidimensional data [29].
Unable to learn the interaction between two
predictors/features under the conditional
independence assumption is the main disadvantage
of NB [2]. Although many rockburst estimation
models have already been described and compared
by previous researchers, developing the accurate
and reliable rockburst predictive model still poses
considerable challenge for burst-prone grounds
[2, 5, 21, 30−36]. Moreover, many other models for
forecasting the rockburst can be considered an
efficient and valuable tool to be applied in other
geological and mining engineering applications.
Particularly, some engineers and scholars are
increasingly interested in combining the output of
several basic classification techniques into one
integrated output using data mining technology,
integrated learning and soft computing methods to
improve classification accuracy [13, 37−40].
However, the integration method has less-
contribution to the rockburst classification than
other fields and requires more extensive
experimental works.
In order to fill the research gap, this paper
investigates a comparative study on the
effectiveness of ensemble learning in rockburst
classification developing two ensemble methods,
i.e., bagging [41], and boosting [42, 43]. Above
methods can be conducted by developing a series of
predictive models which are based on a given
algorithm named the base classifier and vote on all
models in the set to make predictions on new
observations. Classification and regression trees
(CART) conduct by BREIMAN et al [44], is one of
the most favoured algorithms for constructing
classification trees and is usually used as a base
classifier in a classification problem. According to
above discussion, the objective of this investigation
is the contribution to examine the ability of three
CART-based ensemble algorithms for the potential
of rockbursts prediction in hard rock mines. To
accomplish this goal, a research methodology was
developed for the comparison of the performance of
different tree-based ensemble learning algorithms,
including bagging, boosting and CART. These
algorithms were particularly selected due to their
attractions and attentions in various fields of
science and engineering. However, to the best
knowledge of the authors, they have not been
carefully compared with each other for prediction
of rockbursts in hard rock mines. The rest of this
paper includes some explanations regarding
established database and existing empirical models;
then, rockburst prediction in burst-prone mines
using boosting and bagging tree-based ensemble
methods are developed and examined; margin
analysis and the variable relative importance are
also employed to analyze the performance of
proposed models.
2 Database of rockbursts and empirical
models
2.1 Calibration of database
The original database of rockbursts was
reported in the study conducted by ZHOU et al [5]
with 132 case histories and updated by ZHOU et al
[13] with 246 case histories from many kinds of
underground projects (i.e., underground
powerhouse, cavern and tunnel of hydropower
station, railway and road tunnel, coal and hard rock
mines). These databases have been widely-
employed in previous researches [1, 20, 21, 32,
45−47]. With the aim of examining the performance
of the developed tree-based ensemble approaches
for estimating the potential of rockbursts in
burst-prone ground, the data utilized in the present
investigation consists of 102 cases of rockburst
events collected from 14 underground hard rock
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
530
mines with published research works, and is part of
previous database and all the data sources have
been referenced [5, 13]. Field data obtained from
Italian (Raibl lead zinc mine) mine and Russia mine
(Soviet Rasvumchorr workings) besides China hard
rock mines including Baima Iron Mine, Chengchao
Iron Mine, Beiminghe Iron Mine, Maluping
Phosphate Mine, Xincheng Gold Mine, Linlong
Gold Mine, Fankou Lead-Zinc Mine, Jinchuan
Nickel Mine, Tonglushan Copper Mine,
Dongguashan Copper Mine and Hongtoushan
Copper Mine. Out of all the parameters involved in
formulating the present model, the values of
parameters are used directly as available in the
database.
2.2 Data description
The boxplot of the initial dataset is provided in
Figure 2(a). Obviously, the median is not in the
center of the box for most of the data labels, which
means that the distribution of most data labels is not
symmetric (Figure 2(a)). Note that the circles with
blue color indicating outliers. All dependent
indicators have some outliers expect for the uniaxial
compressive strength (UCS or σc) of the rock, the
maximum tangential stress (σθ=MTS) around the
opening and the stress concentration factor
(SCF=σθ/σc) for H rockburst intensity and, the
uniaxial tensile strength (UTS or σt) of the rock and
BI (rock brittleness index (BI=σc/σt)) for L
rockburst intensity, UCS, Wet or EEI (elastic strain
energy index) and SCF for M rockburst intensity,
UTS, SCF and MTS for N rockburst intensity. The
distribution of the rockburst events used in this
work after 1998 is shown in Figure 2(b) as a pie
chart demonstrating the proportion of the four types
of rockburst intensity in underground mines,
categorized as none (N, 17 cases), low (L, 26 cases),
moderate (M, 26 cases) and high (H, 33 cases).
Obviously, this dataset contains a few class
imbalance or sampling bias. The relevant input
indicators used in the development of the rockburst
prediction model with general statistical
characteristics are shown in Figure 2(c) (i.e., range,
mean, standard deviation, median, skew and
kurtosis). Figure 3 shows the scatterplot matrix in
the upper half whereby RPC is the rockburst
prediction classification. The matrix shows
bivariate scatterplots with LOWESS smooth curves
and Pearson correlation values as a side effect [48],
while the two-dimension kernel density estimates
and contours are in the lower half, and the
histograms for each parameter are depicted on the
diagonal. From Figure 3, it is evident that the
parameter MTS is correlated with SCF. In addition,
the joint contours of the kernel density estimation
between these paired indicators is obviously
asymmetric under most circumstances.
2.3 Results on rockburst full data by empirical
criteria methods
Several conventional empirical criteria such as
Russenes criterion [49]; EEI [2, 50], SCF [51], BI
[51, 52], GB50487−2008 and measurable values of
rockburst (MVR) [53] are applied to evaluating the
rockburst potential of 102 case histories. The
applied empirical criteria associated with their
predictive performance on the whole dataset are
presented in Table 1. The predictive accuracy of
above methods is 25.4% (BI by ZHANG et al [52])
to 58.82% (MVR by ZHANG et al [53]) of the
whole dataset and lower than 60% of the filtered
Figure 2 Data description and visualization of combined rockburst database for tree-based ensemble modeling
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
531
Figure 3 Scatterplot matrix of rockburst samples for six parameters
Table 1 Classification results of rockbursts with whole dataset using conventional empirical criteria
Empirical method Equation
Classification criterion Predictive
accuracy/%None Low Moderate High
Russenes criterion[49] SCF=σθ/σc ≤0.2 0.2−0.3 0.3−0.55 >0.55 43.14
Strain energy storage index [50] EEI or Wet 5.0 58.82
Strain energy storage index [2] EEI or Wet 10.0 49.04
Stress concentration factor [2] SCF=σθ/σc 0.7 54.90
GB50487−2008 [2] BI=σc/σt >40 40−26.7 26.7−14.5 7 4−7 2−4 22 25.49
Measurable value of rockburst MVR [53] MVR 0.75 50.00
Note: MVR=tanh{[0.1648 (σθ/σc)3.064 (BI)−0.4625 (Wet)2.672](1/3.6)}.
data. Apparently, the accuracy of rockburst
estimates from empirical criteria is often not
satisfactory due to the influence of various
conditions on rockburst and its complexity of the
features. Fundamentally, these empirical methods
come from their own specific engineering practice,
which leads to the strong dependence of these
methods on the specific engineering background. In
this paper, 102 rockburst cases in 14 mines were
collected, and thus the case engineering types were
extensive, and the locations were quite different.
Moreover, the characteristics of the empirical
method itself determine that it is not a universal
method. Therefore, it can be concluded that the
abovementioned empirical criteria cannot achieve
desirable prediction results from the predictive
accuracy on these rockburst cases.
3 Methodology
3.1 Classification and regression trees (CART)
Recursive partitioning [55] is one of the
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
532
typical decision tree methods. Herein we used
CART which is a form of binary recursive
partitioning [44]. The developing of CART model
comes from the recursive partitioning of dataset: the
constitutive model is defined as a tree structure
whose nodes are associated with data splitting along
one variable [56, 57].
3.2 Bagging
BREIMAN [41] introduced bagging
(Bootstrap aggregating for short) ensemble learning
technology, which can improve classification
accuracy and generalize data patterns. Based on the
training set (Sn), Q bootstrap samples (Sq) are
obtained, where q=1, 2, … , Q. Getting these
bootstrap samples requires drawing the same
number of elements as the original set (n in this
case) to replace. Partial guidance samples may
reject or at least filter partial noise observations,
which means that the classifiers in these sets behave
better than the classifiers constructed in the original
sets. Therefore, bagging is conducive to the
establishment of a better classifier for training sets
with noisy observations. Bagging follows three
simple steps (see Figure 4(a)) [37, 58]:
1) m samples were randomly created from the
training data set. Some slightly different data sets
are allowed to be created in the bootstrapped
samples, but the distribution should be the same as
the entire training set;
2) The classifier built for each bootstrap
sample comes from training an unpruned and single
regression tree;
3) The average prediction value is obtained
before the average individual predictions of these
multiple classifiers.
3.3 Boosting
FREUND and SCHAPIRE [42] imposed a
boosting ensemble learning technology to improve
classification accuracy by transforming a set of
weak classifiers into strong classifiers. Boosting
provides sequential learning of the predictors, and
the major procedure of boosting algorithm is
illustrated in Figure 4(b). In the proposed boosting
algorithm for multi-class classification problems,
two typical boosting algorithms, called
AdaBoost.M1 [37, 42] and SAMME [43, 58], have
been chosen due to their simplicity and natural
extensions.
1) AdaBoost.M1 algorithm
AdaBoost.M1 is based on AdaBoost which
was first tried on multiple classification problems
and has been widely-used [37, 42, 58]. Given a
training set Sn={(x1; y1), …, (xi, yi), …, (xn, yn)}
where y i t akes va lues in 1 , 2 , … , k . Each
Figure 4 Approach to bagging and boosting-based ensemble methodologies: (a) Bagging; (b) Boosting
J. Cent. South Univ. (2021) 28: 527−542
533
observation xi gets the corresponding weight wq(i)
and the initial value is set to 1/n, and then the initial
value is updated with the number of steps.
Constructing a basic classifier Cq(xi) on this new
training set (Sq) and apply it to each training
example. The eq indicates the error of the classifiers
and is calculated as follows [42]:
1
( ) ( ( ) )
n
q q q i i
i
e w i I C x y
(1)
where I(•) is the indicator function which outputs 1
indicating that the internal expression is true,
whereas an output of 0 indicates that the internal
expression is false.
The update of weights requires a constant q to
be calculated by the error of the classifier in the qth
iteration. According to FREUND and SCHAPIRE
[42], q=ln[(1–eq)/eq]. However, BREIMAN [59]
used q=0.5ln[(1–eq)/eq]. Thus, the new weight for
the (q+1)th iteration will be as follows:
wq+1(i)=wq(i)exp[αqI(Cq(xi)≠yi)] (2)
The process step is repeated with the change of
q, q=1, 2, …, Q. The ensemble classifier calculates
the voting weights and values of each class. Then,
the class with the highest vote is assigned.
Specifically,
1
( ) argmax ( ( ) )
Q
f i j Y q q i
q
C x I C x j
(3)
2) SAMME algorithm
SAMME (short for stagewise additive
modeling with loss function of multi-class
exponential), is the second boosting algorithm
implemented herein. Its theoretical development is
first introduced by ZHU et al [43]. Note that, there
is no difference between the SAMME algorithm
and AdaBoost.M1 except for the calculation of the
alpha constant based on the number of classes.
Because of this modification, the SAMME
algorithm only requires that (1−eq)>1/k in order for
the alpha constant to be positive and the weight will
update to follow the right direction [58], that is
αq=ln(k−1)+ln[(1−eq)/eq]. Therefore, the accuracy
of any weak classifier should be higher than the
accuracy of random guess (1/k) rather than the fixed
value 0.5, which is an appropriate requirement for
the two-class case but very demanding for the
multi-class one.
3.4 Margin analysis
The margin of an object is closely related to
the certainty of its classification and is determined
according to the difference between the support of
the true label and the maximum support of the false
label [38, 58, 60]. For the k class, the margin of
each sample xi is obtained by the votes of each class
j in the final ensemble, which are called the support
degree or posterior probability of different classes
μj(xi), j=1, 2, …, k as:
( ) ( ) max ( )i c i j i
j c
m x x x
(4)
where c is the correct class of xi and
1
( ) 1.
k
j i
j
x
Once the data distribution D is determined, the
margin distribution graph (MDG) is defined as the
fraction of instances whose margin is at most λ as a
function of λ∈[−1, 1] [60].
R(λ)=|Dλ|/|D| (5)
where Dλ={x:margin(x)≤λ}, λ∈[−1, 1], |•|
represents size operation, and R(λ)∈[0, 1].
Therefore, negative classified examples could
only appear in all misclassified examples, while
positive classified examples only exist in correctly
classified examples [61]. The margin value for
positive classified observations with high
confidence will approach 1. On the other hand, the
sample margin value with uncertain classification is
small.
3.5 Evaluation criteria
Accuracy is a major concern metric in ML and
the classification accuracy rate (CAR) is taken as
the evaluation criterion of the prediction ability of
the ensemble tree-based ensemble algorithm for
rockburst data [62], which is defined as the
percentage of the amount of casestruly predicted in
the classification model to the total amount of cases,
the main evaluation criterion. The CAR expression
is as follows:
1
1
CAR 100%
C
ii
i
x
n
(6)
4 Tree-based ensemble methods for
rockburst classification
4.1 Indicator analysis
By evaluating the occurring rockburst cases
J. Cent. South Univ. (2021) 28: 527−542
534
and considering their main results and theoretical
methods, a series of single indices or standards for
evaluating the occurrence and intensity of
rockbursts have been put forward to analyze the
rockburst phenomena from multiple angles [2, 5,
13]. Empirical criterion method for rockburst
estimation is primarily used stress/strength,
brittleness, energy and critical depth-based
approaches [2, 5]. The index characteristic of rock
often guides and decides the planning and design of
civil engineering and mining excavation and is also
closely related to the rockburst phenomenon in hard
rock mines [60, 61]. Therefore, it is generally
accepted in many mines to use appropriate rock
index to identify rockburst tendency [62]
particularly UCS and UTS of intact rock are
commonly taken as rockburst potential indicator
[5, 20]. In addition, MTS is another rockburst
indicator because it can reflect the effect of
hydrologic conditions (mainly groundwater), rock
stress environment, opening shape and diameter
factors on rockbursts around the opening [13, 49,
63−67]. Meanwhile, the EEI index defined by
KIDYBINSKIA [51] is another key indicator of
inducing rockburst in hard rock mine. MARTIN
et al [68] mentioned that rockbursts in brittle
hardrock may occur when the SCF reached at some
values. Additionally, the rock brittleness, BI, as
inherent physicomechanical property of rocks
usually, is applied to estimating the potential of
rockbursts [2, 69, 70]. Besides, other engineers
considered the microseismic events for degree of
clustering is proportional with the decrease of
fractal dimension [70]. However, data are needed
on the microseismic event distribution and the
proximity of the events to the seismic source
(hypocenter) which is difficult to be available in our
dataset. Based on these considerations, the input
parameters studied in this paper are the six
indicators (MTS, UCS, UTS, SCF, BI and EEI),
which are also the main indicators for quantitative
assessment of rockburst activities hard rock mines.
Additionally, in previous data-driven researches
equal widely used these indicators [5, 13, 20, 21,
32, 46, 62]. Meanwhile, the potential of rockbursts
is defined using four intensities: none (N), low (L),
medium (M) and high (H). RUSSENES [49] and
ZHOU et al [5, 13] described the classification of
rockbursts in the combined database. In addition,
three different input models are studied which
combine the above indices with the tree-based
ensemble method to evaluate rock burst, and the
traditional parameters of rock burst classification
are selected and labeled ‘√’ in Table 2. Model I of
traditional variables includes MTS, UCS, UTS and
EEI. Model II of traditional variables includes SCF,
BI and EEI. Model III of traditional variables
includes MTS, UCS, UTS, SCF, BI and EEI.
Table 2 Three different models for rockburst assessment
with respect to different input attributes combination
Model number MTS UCS UTS SCF BI Wet
Model I √ √ √ √
Model II √ √ √
Model III √ √ √ √ √ √
4.2 Model development and analysis
This section examines the feasibility of the two
boosting (AdaBoost.M1 and SAMME) and bagging
classification algorithms using the rockburst dataset,
where the CART method is taken as baseline
classifier. Figure 5 demonstrates the basic ML
terms and procedures for building a tree-based
ensemble system. In ML, classifier is used to
predict class of new items which must be tested
after performance in a given dataset. In this regard,
two validation methods were used to guarantee the
comparability and to verify the proposed models
[13, 71]: (a) Train-test split, which means that the
rockburst dataset D is split into two disjointed sets
Dtrain (for training) and Dtest (for testing), and (b)
k-fold cross-validation (CV) which can effectively
avoid over-learning, and the final result is more
convincing than other forms [13]. The original
dataset is randomly classified for a known category
rockburst into two subsets: Dtrain, which is required
to estimate model parameters and construct the each
rockburst classifier model. In this regard, 2/3 of the
available data (68 datasets of all 102 data sets) were
considered to form a training dataset; and Dtest is an
external validation set that tests the performance
and predictive capabilities of each final model. The
test data set contains the remaining 34 data.
Meanwhile, the adabag package [58] is applied and
implemented BREIMAN’s [41] bagging algorithm,
FREUND and SCHAPIRE’s [42] Adaboost.M1
algorithm and ZHU’s [43] Adaboost-SAMME
algori thm with classi f icat ion t rees as base
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
535
Figure 5 Research architectures of tree-based ensemble approach for rockburst prediction
classifiers using the rpart package [72] within the R
environment [73]. Thus, many experiments were
performed in R environment associated with
rockburst data and tree-based ensemble methods.
These algorithms typically use parameter settings
based on default or recommended values.
Furthermore, variable relative importance and
margins analysis are implemented to explain some
characteristics of the ensembles. With the aim of
showing the advantages of using ensembles, results
from a single CART tree as the baseline classifier
are compared to those obtained with two Boostings
and bagging approaches. Considering the number of
rockbursts classes, trees with the maximum depth
of 5 and complexity parameter of −1 are used both
in the single and the ensemble cases, and the
ensembles consist of 250 trees.
With the aim of comparing the error of the four
methods (CART, Bagging, Adaboost.M1 and
SAMME), the mean result is calculated after
10 runs, as described in Table 3. For instance, it can
be seen that the single CART tree model II with a
mean test error of 33.82% is outperformed by
Adaboost.M1 and SAMME, 29.41% and 29.41%,
respectively, and particularly by bagging with a
much lower error (26.47%) maybe because bagging
can reduce the variance error. Similarly, results can
be found in model I and model III. Apparently,
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
536
Table 3 Testing error for 250 iterations with tree models
Method
Model I Model II Model III
Train
set
Test
set
Train
set
Test
set
Train
set
Test
set
CART 0.2794 0.3235 0.2611 0.3382 0.2819 0.3382
Bagging 0 0.2647 0.0312 0.2647 0.0514 0.0882
AdaBoost.M1 0 0.3235 0 0.2941 0 0.1176
SAMME 0 0.3235 0 0.2941 0 0.0882
there is not a significant difference between the
Adaboost.M1 and SAMME methods for the three
models. Taking model III as an example, Figure 6
shows the error development process of training set
and testing set. It is obvious from the figure that the
error values of training set and testing set are
positively correlated with the number of iterations,
and the final error of the test set is less than half of
the error at the beginning of the iteration. In
particular, after the number of iterations reaches the
range 50−100, the error between the training set and
the testing set tends to stabilize. The performance
(CAR) of the test set fell within the range of
67.65% to 91.18% across the twelve models. For
tree-based ensemble classifiers, the confusion
matrix is also a common way to assess theeffectiveness of the proposed models with
independent test set (see Figure 7) [13, 74].
As mentioned above, margin distribution
Figure 7 Confusion matrices and associated rockburst classifier accuracies for tree-based ensemble prediction models
based on independent test set
Figure 6 Models error vs number of trees
for model III: (a) Bagging; (b) Adaboost.
M1; (c) SAMME
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
537
information can be reflected to provide useful
insights and suggestions for ensemble learning
methods with the help of margin analysis [60], and
the margins of these ensemble classifiers were
obtained by the margin () function in the adabag R
package [58]. With the aiming of achiveing the
visualization of results, KUNCHEVA [75]
introduced MDG model to demonstrate the
cumulative distribution of the margins under a
given dataset. Hence, the x-axis is the margin (m) as
defined in Eq. (4) and the y-axis represents the
amount of points where the margin is not greater
than m based on Eq. (5). If the cumulative graph is
a vertical line at m=1, it means that all points are
truly classified and have the greatest possible
certainty. The purpose of MDGs of the training set
(coloured in orange) and the test set (coloured in
blue) is to perform a more detailed margin analysis
of the three ensemble learning methods as described
above. Figure 8 presents the cumulative distribution
of margins for the bagging and two boosting
(AdaBoost.M1 and SAMME) classifiers developed
on the rockburst datasets in this application. In
particular, the cumulative distribution of marginal
value 0 (y-axis) in the MDG represents the
classification error rate. Obvirously, AdaBoost.M1
and SAMME are always positive due to the null
training error as illustrated by the dotted line in
Figure 8(a). It should also be indicated that almost
13% and 8% of the observations in bagging, for
training set and testing set, respectively, achieve a
maximum margin of 1, which is outstanding
considering the large number of classes. This
observation is also true for the results reported in
Table 3. It is hoped that the analysis presented
in this study will provide some interesting
observations and unique insights into the ensemble
learning of marginal analysis.
Overall, the following observations are made
from Table 3 and Figure 8: 1) for the three
tree-based ensemble techniques, the performance
(CAR) of testing set performance fell within the
range of 67.65% to 91.18% across the twelve
models; 2) model III with six input indicators is the
best one when compared with the results from the
two other models with the subsets of attributes; and
3) in the context of determination of rockburst
occurrence and the potential of rockburst, these
methods can be trained to learn the relationships
between the stress condition, rock brittleness and
Figure 8 Margin distribution graph for three models with
bagging, Adaboost.M1 and SAMME methods over
rockburst dataset: (a) Model I; (b) Model II; (c) Model
III
strain energy characteristics with the rockburst
activity, requiring no need to know in advance of
the form of the relationship between them. It is
important to note that these observations are only
based on the dataset used in this work. Whether or
not such observations can be made to other
rockburst datasets is a question that would
necessitate that the new dataset be reanalyzed.
4.3 Relative importance of discriminating
features
Unlike unascertained mathematics, fuzzy
mathematics, extension theory and cloud model, the
developing bagging and boosting tree-based
ensemble techniques do not need to determine the
weight of each index via analytic hierarchy process
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
538
and/or information entropy. It is worth noting that
bagging and boosting ensemble learning methods
can provide relative importance of predictive
variables based on their predictability of rockbursts
potential. For this goal, the measure of importance
considers not only the Gini yield given by a feature
in each tree, but also the weights of trees in the
bagging, AdaBoost.M1 and SAMME cases.
Figure 9 shows the variables ranked by its relative
importance value for each rockburst model with
Figure 9 Variables relative importance for each model:
(a) Model I; (b) Model II; (c) Model III
bagging and the two boosting ensemble classifiers.
In terms of model I (Figure 9(a)), the most sensitive
factor in the indicators is EEI, followed by the
indicators MTS, UTS, UCS; In terms of model II
(Figure 9(b)), EEI has the greatest impact on results
when making the same changes compared to other
indicators, followed by the indicators SCF, BI; In
terms of model III (Figure 9(c)), EEI is still the
most sensitive factor among the indicators,
followed by the indicators MTS, SCF, UTS, UCS,
and BI. Findings demonstrate that among the
rockburst classification prediction indicators, EEI is
also the most relevant predictor. Destress blasting
techniques commonly used to avoid high-stress
concentrations around underground openings have
been proven very effective to mitigate the risks of
rockburst [76].
4.4 Limitations
There are some limitations regarding this work
that will be perfected in future research. First, the
proposed ensemble method is designed for general
purpose but is only validated on 102 datasets.
Additional datasets can be used for further
validation. Secondly, class imbalance and sampling
bias are two representative important factors of
model uncertainty that have an impact on the
predictive probability of rockburst classification
models and should be addressed in future studies.
Thirdly, tuning hyper-parameters of ensemble
models and combining more baseline classifier
models are not addressed for further improvement
on the performance of ensemble techniques in this
work. Meanwhile, the rockburst is associated with
in-situ stress, rock property, structure of rock mass,
and dynamic disturbance effect on rockbursts [2, 4,
63, 64]. Thus it is very difficult to predict time-
space distribution of rockburst exactly. Moreover,
note that only strainburst type in hard rock mines is
developed in this work, other types of rockburst
such as fault-slip are not addressed due to data
limitation. Also, other additional indicators such as
peak particle velocity (PPV) [60, 61] at the time of
rockburst and local geological joint or fractures
[62, 77] may be applied to improving the
performance of rockburst potential prediction;
however, collecting such data can be cumbersome.
Finally, the “black box” feature of the ensemble
algorithm is a dark cloud that hangs over the
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
539
complex relationship between the interpreted
response and the predictor variable.
5 Conclusions
Models for forecasting rockburst with high
accuracy can be beneficial tools for mitigating the
risk of rockburst hazards and strengthening mine
safety in burst-prone grounds. In this work, six
traditional rockburst indicators were measured and
three CART-based models for ensemble approaches
were presented and compared for the prediction of
rockburst in hard rock mines can be beneficial tools
for work in burst-prone grounds. A data set of 102
rockburst cases compiled from 14 hard rock mines
is applied to developing the proposed models.
Findings reveal that the performance of the
proposed models with combination of the indicators
with ensemble methods works much better than that
with eight empirical criteria. Bagging was
particularly found to be the best model among all
three ensemble methods, while thereis not a big
difference between the two boosting methods from
the aspect of accurancy. These results illustrated
that ensemble learning approaches can be taken as a
viable tool for identifying rockburst. Meanwhile,
bagging and two boosting methods demonstrate that
the EEI index is the most relevant indicator among
the features for the task of rockburst classification
prediction. In summary, the main work of this paper
was to provide a relatively reliable model for
rockburst prediction. To further enhance the
accuracy of prediction models, the authors hope to
fuse the results obtained from different empirical
criteria and ML methods in future work with more
rockburst case histories.
Contributors
WANG Shi-ming and ZHOU Jian provided the
concept and wrote the original draft of manuscript.
ZHOU Jian, LI Chuan-qi and Danial Jahed
ARMAGHANI collected, visualized and analyzed
dataset, conducted the models. LI Xi-bing and Hani
S. MITRI edited the draft of manuscript.
Conflict of interest
The authors declare that they have no conflicts
of interest.
References
[1] LI X, ZHOU J, WANG S, LIU B. Review and practice of
deep mining for solid mineral resources [J]. Chinese Journal
of Nonferrous Metals, 2017, 27(6): 1236−1262. DOI:
10.19476/j. ysxb.1004.0609.2017. 06.021. (in Chinese)
[2] ZHOU J, LI X, MITRI H S. Evaluation method of rockburst:
State-of-the-art literature review [J]. Tunn Undergr Sp Tech,
2018, 81: 632−659. DOI: 10.1016/j.tust.2018.08.029.
[3] LI X, GONG F, TAO M, DONG L, DU K, MA C, ZHOU Z,
YIN T. Failure mechanism and coupled static-dynamic
loading theory in deep hard rock mining: A review [J].
Journal of Rock Mechanics and Geotechnical Engineering,
2017, 9(4): 767−782. DOI: 10.1016/j.jrmge.2017.04.004.
[4] KAISER P K, MCCREATH D R, TANNANT D D.
Canadian rockburst research program 1990−1995, vols 1−6
[R]. Sudbury, Ontario: CAMIRO Mining Division, 1997.
[5] ZHOU J, LI X B, SHI X Z. Long-term prediction model of
rockburst in underground openings using heuristic
algorithms and support vector machines [J]. Safety Science,
2012, 50(4): 629−644. DOI: 10.1016/j.ssci.2011.08.065.
[6] GONG F Q, SI X F, LI X B, WANG S Y. Experimental
investigation of strain rockburst in circular caverns under
deep three-dimensional high-stress conditions [J]. Rock
Mechanics and Rock Engineering, 2019, 52(5): 1459−1474.
DOI: 10.1007/s00603-018-1660-5.
[7] SI X, GONG F. Strength-weakening effect and shear-tension
failure mode transformation mechanism of rockburst for
fine-grained granite under triaxial unloading compression [J].
International Journal of Rock Mechanics and Mining
Sciences, 2020, 131: 104347. DOI: 10.1016/j.ijrmms.2020.
104347.
[8] SHI X Z, ZHOU J, DONG L, HU H Y, WANG H Y, CHEN
S R. Application of unascertained measurement model to
prediction of classification of rockburst intensity [J]. Chinese
Journal of Rock Mechanics and Engineering, 2010, 29(s1):
2720−2727. (in Chinese)
[9] ORTLEPP W D. RaSiM comes of age—A review of the
contribution to the understanding and control of mine
rockbursts [C]// POTVIN Y, HUDYMA M. Controlling
Seismic Risk — Proceedings of Sixth International
Symposium on Rockburst and Seismicity in Mines.
Australian Centre for Geomechanics, 2005: 3−20.
[10] WENG L, HUANG L Q, TAHERI A, LI X B. Rockburst
characteristics and numerical simulation based on a strain
energy density index: A case study of a roadway in linglong
gold mine, China [J]. Tunnelling & Underground Space
Technology, 2017, 69: 223−232. DOI: 10.1016/j.tust.2017.
05.011.
[11] JIANG Q, FENG X T, XIANG T B, SU G S. Rockburst
characteristics and numerical simulation based on a new
energy index: A case study of a tunnel at 2,500 m depth [J].
Bulletin of Engineering Geology and the Environment, 2010,
69(3): 381−388. DOI: 10.1007/s10064-010-0275-1.
[12] ZHU G A, DOU L M, CAO A Y, CAI W, WANG C B, LIU
Z G, LI J. Assessment and analysis of strata movement with
special reference to rock burst mechanism in island longwall
panel [J]. Journal of Central South University, 2017, 24(12):
2951−2960. DOI: 10.1007/s11771-017-3709-0.
francisco.wong
Resaltado
francisco.wong
Resaltado
francisco.wong
Resaltado
J. Cent. South Univ. (2021) 28: 527−542
540
[13] ZHOU J, LI X, MITRI H S. Classification of rockburst in
underground projects: Comparison of ten supervised learning
methods [J]. Journal of Computing in Civil Engineering,
2016, 30(5): 04016003. DOI: 10.1061/(ASCE)CP.1943-5487.
0000553.
[14] LUO Y, GONG F Q, LI X B, WANG S Y. Experimental
simulation investigation of influence of depth on spalling
characteristics in circular hard rock tunnel [J]. Journal of
Central South University, 2020, 27(3): 891−910. DOI:
10.1007/S11771-020-4 339-5.
[15] MITRI H S, TANG B, SIMON R. FE modelling of
mining-induced energy release and storage rates [J]. The
Journal of the South African Institute of Mining and
Metallurgy, 1999, 99(2): 103−110.
[16] GONG F, YAN J, LI X, LUO S. A peak-strength strain
energy storage index for rock burst proneness of rock
materials [J]. International Journal of Rock Mechanics and
Mining Sciences, 2019, 117: 76−89. DOI: 10.1016/
j.ijrmms.2019.03.020
[17] FENG X T, WANG L N. Rockburst prediction based on
neural networks [J]. Transactions of Nonferrous Metals
Society of China, 1994, 4(1): 7−14.
[18] GONG F Q, LI X B. A distance discriminant analysis
method for prediction of possibility and classification of
rockburst and its application [J]. Chinese Journal of Rock
Mechanics and Engineering, 2007, 26(5): 1012−1018. (in
Chinese)
[19] ZHOU J, SHI X Z, DONG L, HU H Y, WANG H Y. Fisher
discriminant analysis model and its application for prediction
of classification of rockburst in deep buried long tunnel [J].
Journal of Coal Science and Engineering, 2010, 16(2):
144−149. (in Chinese)
[20] LIU Z, SHAO J, XU W, MENG Y. Prediction of rock burst
classification using the technique of cloud models with
attribution weight [J]. Natural Hazards, 2013, 2: 549−568.
DOI: 10.1007/s11 069-013-0635-9.
[21] ADOKO A C, GOKCEOGLU C, WU L, ZUO Q J.
Knowledge-based and data-driven fuzzy modeling for
rockburst prediction [J]. International Journal of Rock
Mechanics and Mining Sciences, 2013, 61: 86−95. DOI:
10.1016/j.ijrmms.2013.02.010.
[22] ZHOU J, SHI X Z, HUANG R D, QIU X Y, CHEN C.
Feasibility of stochastic gradient boosting approach for
predicting rockburst damage in burst-prone mines [J]. Trans
Nonferrous Meterol Soc China, 2016, 26: 1938−1945. DOI:
10.1016 /S1003-6326(16)64312-1.
[23] ZHOU J, GUO H, KOOPIALIPOOR M, ARMAGHANI D J,
TAHIR M M. Investigating the effective parameters on the
risk levels of rockburst phenomena by developing a hybrid
heuristic algorithm [J]. Engineering with Computers, 2020:
1−16. DOI: 10.1007/s00366-019-00908-9.
[24] LE L T, NGUYEN H, DOU J, ZHOU J. A comparative
study of PSO-ANN, GA-ANN, ICA-ANN, and ABC-ANN
in estimating the heating load of buildings’ energy efficiency
for smart city planning [J]. Applied Sciences, 2019, 9(13):
2630. DOI:10.3390/app9132630.
[25] ZHOU J, LI C, ARSLAN C A, HASANIPANAH M,
AMNIEH H B. Performance evaluation of hybrid
FFA-ANFIS and GA-ANFIS models to predict particle size
distribution of a muck-pile after blasting [J]. Engineering
with Computers, 2019. DOI: 10.1007/s00366-019-00822-0.
[26] ZHOU J, NEKOUIE A, ARSLAN C A, PHAM B T,
HASANIPANAH M. Novel approach for forecasting the
blast-induced AOp using a hybrid fuzzy system and firefly
algorithm [J]. Engineering with Computers, 2020, 36(2):
703−712. DOI: 10. 1007/s00366-019-00725-0.
[27] MARZI H, HAJ DARWISH A, HELFAWI H. Training
ANFIS using the enhanced Bees Algorithm and least squares
estimation [J]. Intelligent Automation & Soft Computing,
2017, 23(2): 227−234. DOI: 10.1080/10798587.2016.119
6880.
[28] YU Z, SHI X, ZHOU J, CHEN X, MIAO X, TENG B,IPANGELWA T. Prediction of blast-induced rock
movement during bench blasting: Use of gray wolf optimizer
and support vector regression [J]. Natural Resources
Research, 2020, 29(2): 843−65. DOI: 10.1007/s11053-
019-09593-3.
[29] ZHANG S. Cost-sensitive KNN classification [J].
Neurocomputing, 2020, 391: 234−242. DOI: 10.1016/j.neuc
om.2018.11.101.
[30] HE M, XIA H, JIA X, GONG W, ZHAO F, LIANG K.
Studies on classification, criteria and control of rockbursts
[J]. Journal of Rock Mechanics and Geotechnical
Engineering, 2012, 4(2): 97−114. DOI: 10.3724/SP.J.
1235.2012.00097.
[31] CAI W, DOU L, ZHANG M, CAO W, SHI J Q, FENG L. A
fuzzy comprehensive evaluation methodology for rock burst
forecasting using microseismi monitoring [J]. Tunnelling and
Underground Space Technology, 2018, 80: 232−245. DOI:
10.1016/j.tust.2018.06.029.
[32] AFRAEI S, SHAHRIAR K, MADANI S H. Developing
intelligent classification models for rock burst prediction
after recognizing significant predictor variables, Section 1:
Literature review and data preprocessing procedure [J].
Tunnelling and Underground Space Technology, 2019, 83:
324−353.
[33] GONG F Q, LI X B, ZHANG W. Rockburst prediction of
underground engineering based on Bayes discriminant
analysis method [J]. Rock and Soil Mechanics, 2010,
31(Suppl 1): 370−377. DOI: 10.1109/VETECS.2006.
1683241. (in Chinese)
[34] LI T Z, LI Y X, YANG X L. Rock burst prediction based on
genetic algorithms and extreme learning machine [J]. Journal
of Central South University, 2017, 24(9): 2105−2113. DOI:
10.1007/s 11771-017-3619-1.
[35] ZHOU J, KOOPIALIPOOR M, LI E, ARMAGHANI D J.
Prediction of rockburst risk in underground projects
developing a neuro-bee intelligent system [J]. Bulletin of
Engineering Geology and the Environment, 2020, 5: 1−15.
DOI: 10.1007/s10064-020-01788-w.
[36] WANG C, WU A, LU H, BAO T, LIU X. Predicting
rockburst tendency based on fuzzy matter–element model [J].
International Journal of Rock Mechanics and Mining
Sciences, 2015, 75: 224−232. DOI: 10.1016/j.ijrmms.2015.
J. Cent. South Univ. (2021) 28: 527−542
541
02.004.
[37] HASTIE T, TIBSHIRANI R, FRIEDMAN J. The elements
of statistical learning: Data mining, inference, and prediction
[M]. 2nd ed. New York: Springer-Verlag, 2009.
[38] KIM M J, KANG D K. Ensemble with neural networks for
bankruptcy prediction [J]. Expert Systems with Applications,
2010, 37(4): 3373−3379. DOI: 10.1016/j.eswa.2009.10.012.
[39] CHOU J S, CHIEH L C. Predicting disputes in public-
private partnership projects: Classification and ensemble
models [J]. Journal of Computing in Civil Engineering, 2013,
27(1): 51−60. DOI: 10.1061/(ASCE)CP.1943-5487.0000197.
[40] ZHOU J, SHI X, DU K, QIU X, LI X, MITRI H S.
Feasibility of random-forest approach for prediction of
ground settlements induced by the construction of a
shield-driven tunnel [J]. International Journal of
Geomechanics, 2017, 17(6): 04016129. DOI: 10.1061/
(ASCE)GM.1943-5622.0000817.
[41] BREIMAN L. Bagging predictors [J]. Machine Learning,
1996, 24(2): 123−140. DOI: 10.1007 /BF00058655.
[42] FREUND Y, SCHAPIRE R E. Experiments with a new
boosting algorithm [C]// Proceedings of the Thirteenth
International Conference on Machine Learning. Morgan
Kaufmann, 1996: 148−156.
[43] ZHU J, ZOU H, ROSSEZ S, HASTIE T. Multi-class
AdaBoost [J]. Statistics and its Interface, 2009, 2: 349−360.
DOI: 10.4310/SII.2009.v2.n3.a8.
[44] BREIMAN L, FRIEDMAN J H, OLSHEN R A, STONE C J.
Classification and regression trees [M]. Wadsworth, Belmont,
1984.
[45] ALFARO E, GARCIA N, GAMEZ M, ELIZONDO D.
Bankruptcy forecasting: An empirical comparison of
AdaBoost and neural networks [J]. Decision Support
Systems, 2008, 45: 110−122. DOI: 10.1016/j.dss.2007.12.
002.
[46] FARADONBEH R S, TAHERI A. Long-term prediction of
rockburst hazard in deep underground openings using three
robust data mining techniques [J]. Engineering with
Computers, 2018, 35(2): 659−675. DOI: 10.1007/s00366-
018-0624-4.
[47] GHASEMI E, GHOLIZADEH H, ADOKO A C. Evaluation
of rockburst occurrence and intensity in underground
structures using decision tree approach [J]. Engineering with
Computers, 2020, 36(1): 213−225. DOI: 10.1007/s00366-
018-00695-9.
[48] CLEVELAND W S. LOWESS: A program for smoothing
scatterplots by robust locally weighted regression [J]. The
American Statistician, 1981, 35: 54.
[49] RUSSENES B F. Analysis of rock spalling for tunnels in
steep valley sides (in Norwegian) [D]. Trondheim:
Norwegian Institute of Technology, 1974.
[50] KIDYBINSKI A. Bursting liability indices of coal [J].
International Journal of Rock Mechanics and Mining
Sciences and Geomechanics Abstracts, 1981, 18(4): 295−304.
DOI: 10.1016/0148-9062 (81) 91194-3.
[51] WANG Y H, LI W D, LI Q G, XU Y, TAN G H. Method of
fuzzy comprehensive evaluations for rockburst prediction [J].
Chinese Journal of Rock Mechanics and Engineering, 1998,
17(5): 493−501. (in Chinese)
[52] ZHANG J J, FU B J, LI Z K, SONG S W, SHANG Y J.
Criterion and classification for strain mode rockbursts based
on five-factor comprehensive method [C]// Proceedings of
the 12th ISRM International Congress on Rock Mechanics,
Harmonising Rock Engineering and the Environment.
London: Taylor & Francis Group, 2011: 1435−1440.
[53] ZHANG G C, GAO Q, DU J Q, LI K K. Rockburst criterion
based on artificial neural networks and nonlinear regression
[J]. Journal of Central South University (Science and
Technology), 2013, 44(7): 2977−2981. (in Chinese)
[54] PENG Z, WANG Y H, LI T J. Griffith theory and rock burst
of criterion [J]. Chinese Journal of Rock Mechanics and
Engineering, 1996, 15(Suppl): 491−495. (in Chinese)
[55] HOTHORN T, KURT HORNIK K, ZEILEIS A. Unbiased
recursive partitioning: A conditional inference framework [J].
Journal of Computational and Graphical Statistics, 2006,
15(3): 651−674. DOI: 10.1198/106186006X133933.
[56] VENABLES W N, RIPLEY B D. Modern applied statistics
with S [M]. Springer, 2002.
[57] GARZÓN M B, BLAZEK R, NETELER M, de DIOS R S,
OLLERO H S, FURLANELLO C. Predicting habitat
suitability with machine learning models: The potential area
of Pinus Sylvestris L. in the Iberian Peninsula [J]. Ecol
Model, 2006, 97: 383−393. DOI: 10.1016/j.ecolmodel.2006.
03.015.
[58] ALFARO E, GAMEZ M, GARCIA N. Adabag: An R
package for classification with boosting and bagging [J].
Journal of Statistical Software, 2013, 54(2): 1−35. DOI:
10.18637/jss.v054.i02.
[59] BREIMAN L. Arcing classifiers [J]. The Annals of Statistics
1998, 26(3): 801−849. DOI: 10.1214 /aos/1024691079.
[60] SCHAPIRE R E, FREUND Y, BARTLETT P, LEE W S.
Boosting the margin: A new explanation for the effectiveness
of voting methods [J]. The Annals of Statistics, 1998, 26(5):
1651−1686.
[61] HE H, CAO Y. SSC: A classifier combination method based
on signal strength [J]. IEEE Transactions on Neural
Networks & Learning Systems, 2012, 23(7): 1100. DOI:
10.1109/TNNL S.2012.2198227.
[62] GRINAND C, ARROUAYS D, LAROCHE B, MARTIN
MP. Extrapolating regional soil landscapes from an existing
soil map: Sampling intensity, validation procedures, and
integration of spatial context [J]. Geoderma, 2008, 143:
180−190. DOI: 10.1016/j.geoderma.2007.11.004.
[63] GONG F Q, LUO Y, LI X B, SI X F, TAO M. Experimental
simulation investigation on rockburst induced by spalling
failure in deep circular tunnels [J]. Tunnelling and
Underground Space Technology, 2018, 81: 413−427. DOI:
10.1016/j.tust.2018.07.035.
[64] TAO M, ZHAO H, LI X, LI X, DU K. Failure characteristics
and stress distribution of pre-stressed rock specimen with
circular cavity subjected to dynamic loading [J]. Tunnelling
and Underground Space Technology, 2018, 81: 1−15. DOI:
10.1016/j.tust.2018.06.028.
[65] SINGH S P. Classification of mine workings according to
their rockburstproneness [J]. Mining Science and
Technology, 1989, 8(3): 253−262. DOI: 10.1016/S0167-
9031(89)90404-0.
[66] BARTON N, LIEN R, LUNDE J. Engineering classification
J. Cent. South Univ. (2021) 28: 527−542
542
of rock masses for the design of tunnel support [J]. Rock
Mechanics and Rock Engineering, 1974, 6(4): 189−236. DOI:
10.1007/bf01239496.
[67] HOEK E, BROWN E T. Underground excavation in rock
[M]. London: The Institute of Mining and Metallurgy, 1980.
[68] MARTIN C, KAISER P, MCCREATH D. Hoek-Brown
parameters for predicting the depth of brittle failure around
tunnels [J]. Canadian Geotechnical Journal, 1999, 36:
136−151. DOI: 10.1139/cgj-36-1-136.
[69] SINGH S P. The influence of rock properties on the
occurrence and control of rockbursts [J]. Mining Science and
Technology, 1987(5): 11−18. DOI: 10.1016/S0167-9031(87)
90854-1.
[70] XIE H. PARISEAU W G. Fractal character and the
mechanism of rock bursts [J]. International Journal of Rock
Mechanics and Mining Sciences & Geomechanics Abstracts,
1993, 30(4): 343−350. DOI: 10.1016/0148-9062(93)91
718-X.
[71] ZHOU J, QIU Y, ZHU S, ARMAGHANI DJ, LI C,
NGUYEN H, YAGIZ S. Optimization of support vector
machine through the use of metaheuristic algorithms in
forecasting TBM advance rate [J]. Engineering Applications
of Artificial Intelligence, 2021, 97: 104015.
[72] THERNEAU T, ATKINSON E. An introduction to recursive
partioning using the rpart routine [R]. Rochester: Section of
Biostatistics, 1997.
[73] ROSS I, ROBERT G. R: A language and environment for
statistical computing. R foundation for statistical computing
[J]. Journal of Computational and Graphical Statistics, 2008,
5: 299−314. DOI: 10.1890/0012-9658 (2002)083 [3097:
CFHIWS]2.0.CO;2.
[74] ZHOU J, LI X, MITRI H S. Comparative performance of six
supervised learning methods for the development of models
of hard rock pillar stability prediction [J]. Natural Hazards,
2015, 79(1): 291−316. DOI: 10.1007/s11069-015-1842-3.
[75] KUNCHEVA L I. Combining pattern classifiers: Methods
and algorithms [M]. New York: Wiley, 2004.
[76] MITRI H S. Practitioner’s guide to destress blasting in hard
rock mines [R]. Montreal: McGill University, 2000.
[77] JIANG Q, YANG B, YAN F, LIU C, SHI Y, LI L. New
method for characterizing the shear damage of natural rock
joint based on 3D engraving and 3D scanning [J].
International Journal of Geomechanics, 2020, 20(2):
06019022. DOI: 10.1061/(ASCE) GM.1943-5622.0001575.
(Edited by YANG Hua)
中文导读
用基于树的 Bagging 和 Boosting 集成技术预测硬岩矿山岩爆
摘要:岩爆预测对地下硬岩矿山的设计和施工至关重要。使用三种基于树的集成方法,对由 102 个历
史案例(即 1998—2011 年期间 14 个硬岩矿山数据)组成的岩爆数据库进行了检查,以用于有岩爆倾向
矿井的岩爆预测。该岩爆数据集包含六个广泛接受的倾向性指标,即:开挖边界周围的最大切向应力
(MTS)、完整岩石的单轴抗压强度(UCS)和单轴抗拉强度(UTS)、应力集中系数(SCF)、岩石脆性指数(BI)
和应变能储存指数(EEI)。以分类树作为基准分类器的两种 Boosting 算法(AdaBoost.M1,SAMME)和
Bagging 算法进行了评估,评估了它们学习岩爆的能力。将可用数据集随机分为训练集(整个数据集的
2/3)和测试集(其余数据集)。采用重复 10 倍交叉验证(CV)作为调整模型超参数的验证方法,并利用边
际分析和变量相对重要性分析了各集成学习模型特征。根据重复 10 倍交叉验证结果,对岩爆数据集
的精度分析表明,与 AdaBoost.M1、SAMME 算法和岩爆经验判据相比,Bagging 方法是预测硬岩矿
山岩爆的最佳方法。
关键词:岩爆;硬岩;预测;Bagging 方法;Boosting 方法;集成学习