
Pearson (1985) gave a four-parameter system of probability density functions and fitted the parameters by the method of moments (MME). Tukey (1960) proposed one-parameter lambda distribution. Tukey’s lambda was generalized, for the purpose of generating random variables for Monte Carlo simulation studies, to the four-parameter generalized lambda distribution (GLD), by Ramberg and Schmeiser (1972) and Ramberg and Schmeiser (1974). Ramberg
where
where 0 ≤
More details for GLD family refer to Karian and Dudewicz (2000). This paper is organized as follows. In Section 2, we develop the change point detection procedure for the GLD model based on the MIC and provides preliminaries on how to construct a confidence curve and confidence set for the specified level. The simulation results are presented in Section 3. Our new method for the goodness-of-fit of the asymptotic null distribution of
In this section, we use the modified information approach (MIC) to detect changes in a GLD model. Chen
In general, we consider multiple changes in the data set. Vostrikova (1981) proposed the binary segmentation method which could detect multiple structural changes recursively at most one change point at each step. In the first step, this method allows us to scan the whole data set by testing the null hypothesis of no change versus the alternative hypothesis of having one change. Once the first change (if there is any) has been located, the data is divided to two subsequences which are before the change point and after the change point. Then the second step is to repeat the same scanning procedure in Step 1 to these two subsequences respectively by assuming at most one change in each subsequence. Such a process will be repeated until there will be no further subsequences having change points. By doing so, we can find all the possible changes as well as estimate their locations. She also showed that the binary segmentation procedure is consistent. In particular, the multiple change point problem may be viewed as a single problem along with the binary segmentation procedure. Therefore, in this paper, we establish the testing procedure for a single change point detection method, however, it can be easily generalised to multiple change point problem, if needed.
Let
We are interested in testing the changes in all parameters simultaneously. Thus, we would like to test the hypotheses are,
versus,
where 1 ≤
where
which indicates there is no change point in the data set, and we reject
which indicates that there exists a change point in the data set. Consequently we can estimate the change point location
Further, we define the test statistic
We reject the null hypothesis for a sufficiently large value of
Confidence distributions (CD) are distribution estimates to be interpreted as distributions of epistemic probabilities. The concept of a CD is similar to a point estimator and it can be referred to as a sample-dependent distribution that can represent confidence intervals of all levels for a parameter of interest. A formal definition of CD can be found in Schweder and Hjort (2002). Furthermore, Schweder and Hjort (2016) systematically studied the theoretical properties of the CD. A detailed analysis of recent developments of CD has been given by Xie and Singh (2013). More applications of the CD can be found in the literature, including bootstrap distributions,
The CD for change point analysis has been investigated by Cunen
Suppose
where Θ
where Θ̂
where
where the
for a large number of
In this section, due to the difficulty in deriving its the analytic properties, we use simulations to investigate the critical values and the performance of the proposed test statistic
We now describe how to obtain empirical critical values for our test statistic
where BIC(
The major difference between
There are two general approaches available to compute critical values. They are simulation-and bootstrap-based approaches. The simulation-based approach requires the estimation of the test statistics values
When using the bootstrap method to obtain simulated critical values of a test statistic, we need to ensure that the bootstrap samples are re-sampled from data under the null distribution. In the simulation-based approach, however, the distribution under the null hypothesis has been determined before re-sampling. Therefore, it is known to satisfy
where
In our simulation study, we set up the null distribution to be GLD(2, 1, 0.19, 0.19) and choose sample sizes
Step 1: We generate data with various sample sizes
Step 2: For each generated sample, we calculate
Step 3: We repeat the above steps
The empirical critical values are provided in Table 1.
We should note here that, for the real data application, we should follow the bootstrap method proposed in Section 3.1 to calculate the
In this subsection, we provide results of the simulation study for the coverage probability, confidence sets, and consistency of the change point estimator. Three different sample sizes
In this section, we conducted simulations under different scenarios to investigate the performance of test procedures based
Further, various sample sizes
Regardless of the method used for the power calculations, the power of the test increases as the sample size becomes larger. Moreover, it appears that the MIC based method gives larger power compared to the BIC based method. This may be due to that MIC method depends on the location of the change point. We also notice that the power of the test increases as the differences between the parameters increases.
First, we examine the coverage probabilities when the method has the exact right coverage for the specified level. We compare our method, the MIC based method, with the log-likelihood based approach proposed in Cunen
Next, we compute the average sizes of confidence sets for MIC and log-likelihood based methods. The results are summarized in Table 4. It can be seen that the MIC based method gives smaller confidence sets compared to the log-likelihood based approach proposed in Cunen
We also investigate the consistency of the estimator
According to Chen
Energy distance, as described in Székely (2000), is defined to be the statistical distance between probability distributions. The associated statistics, named energy statistics, are the function of energy distances. The concept is motivated by Newton’s gravitational potential energy which is a function of the distance between two objects. Thus the idea of energy statistics is to consider statistical observations as heavenly bodies governed by a statistical potential energy, which is zero if and only if an underlying statistical null hypothesis is true. Székely and Rizzo (2013) defined the energy distance ℰ(
provided
Then the one-sample energy statistic for the goodness-of-fit test based on the energy distance is given by the following definition.
As we mentioned earlier, according to Chen
The proof of Theorem 3 is similar to Rizzo and Haman (2016). Thus, details are omitted to conserve space. According to Rizzo (2002), the last term of the (
where
where
Generate a data
Compute the energy statistics of the data
Repeat Steps 1 & 2 for 5000 times and obtain
The critical value can be obtained by finding a 95% quantile of the energy statistics.
Simulate
Compute the energy statistic of the data
Compare the energy statistic in Step 6 with the critical value found in Step 4, if the critical value exceeds the energy statistic, we conclude that
We follow the above procedure to conduct the one sample energy goodness-of-fit test statistic. The critical value at 5% significance is 9.5037. The energy statistic for the
Figure 1 shows that there is a significant deviation from the reference line. This confirms our previous results, that the asymptotic null distribution of
In this section, the proposed method is applied to analyze two stock market returns data from the Brazilian and Chilean markets. These data sets were previously used in the literature Ngunkeng and Ning (2014), Ratnasingam and Ning (2020). We assume that changes occur simultaneously across all four parameters. The stock return ratio is obtained through the following transformation.
In order to identify multiple changes and to create the appropriate confidence sets, we use the proposal method along with binary segmentation procedure. First, MIC(262) = −807.3228. Then the min1≤
First, we compute MIC(
In this paper, we propose a change point detection procedure for a GLD model based on the modified information criterion. In order to use as much information about the change point location, the proposed procedure takes into account the effect in terms of model complexity in regards to the location of the change point. We provide confidence sets for the change point location for a specified level
Critical values of
Method | ||||
---|---|---|---|---|
0.01 | 0.05 | 0.1 | ||
50 | 13.2672 | 15.6640 | 19.0798 | |
9.3552 | 11.7520 | 15.1678 | ||
60 | 12.5919 | 14.5480 | 18.9868 | |
8.4976 | 10.4537 | 14.8925 | ||
80 | 11.2876 | 13.1865 | 18.5648 | |
6.9056 | 8.8045 | 14.1827 | ||
100 | 11.4198 | 13.6665 | 17.8261 | |
6.8146 | 9.0613 | 13.2209 | ||
150 | 9.6414 | 11.5773 | 15.7964 | |
4.6307 | 6.5666 | 10.7858 |
Power comparison between MIC and BIC for
Model | ( | ||||
---|---|---|---|---|---|
(2.5, 1.5, 0.69, 0.69) | (3, 2, 1.19, 1.19) | (4, 3, 2.19, 2.19) | |||
50 | 10 | MIC | 0.796 | 0.882 | 0.930 |
BIC | 0.726 | 0.874 | 0.926 | ||
15 | MIC | 0.862 | 0.930 | 0.960 | |
BIC | 0.840 | 0.924 | 0.950 | ||
25 | MIC | 0.952 | 0.988 | 0.976 | |
BIC | 0.934 | 0.988 | 0.972 | ||
100 | 15 | MIC | 0.892 | 0.972 | 0.998 |
BIC | 0.882 | 0.968 | 0.996 | ||
25 | MIC | 0.914 | 0.982 | 1.000 | |
BIC | 0.908 | 0.978 | 0.998 | ||
50 | MIC | 0.972 | 0.992 | 1.000 | |
BIC | 0.964 | 0.992 | 1.000 | ||
150 | 35 | MIC | 0.948 | 0.980 | 1.000 |
BIC | 0.938 | 0.978 | 1.000 | ||
50 | MIC | 0.974 | 0.998 | 1.000 | |
BIC | 0.972 | 0.998 | 1.000 | ||
75 | MIC | 0.994 | 1.000 | 1.000 | |
BIC | 0.986 | 1.000 | 1.000 |
Coverage probability comparison between MIC and log-likelihood methods
( | ||||||||
---|---|---|---|---|---|---|---|---|
(2.5, 1.5, 0.69,0.69) | (3,2, 1.19,1.19) | (4,3,2.19,2.19) | ||||||
MIC | loglik | MIC | loglik | MIC | loglik | |||
50 | 10 | 0.50 | 0.38 | 0.35 | 0.40 | 0.38 | 0.45 | 0.42 |
0.90 | 0.80 | 0.78 | 0.88 | 0.87 | 0.90 | 0.88 | ||
0.95 | 0.86 | 0.84 | 0.93 | 0.92 | 0.95 | 0.95 | ||
0.99 | 0.98 | 0.96 | 0.99 | 0.98 | 0.99 | 1.00 | ||
15 | 0.50 | 0.42 | 0.39 | 0.42 | 0.40 | 0.52 | 0.52 | |
0.90 | 0.82 | 0.82 | 0.89 | 0.88 | 0.91 | 0.90 | ||
0.95 | 0.91 | 0.89 | 0.95 | 0.94 | 0.95 | 0.95 | ||
0.99 | 0.98 | 0.98 | 0.99 | 0.98 | 1.00 | 0.99 | ||
25 | 0.50 | 0.46 | 0.44 | 0.44 | 0.43 | 0.52 | 0.52 | |
0.90 | 0.88 | 0.87 | 0.89 | 0.89 | 0.89 | 0.89 | ||
0.95 | 0.93 | 0.92 | 0.95 | 0.94 | 0.95 | 0.95 | ||
0.99 | 0.98 | 0.98 | 1.00 | 0.99 | 0.99 | 1.00 | ||
100 | 15 | 0.50 | 0.44 | 0.42 | 0.49 | 0.48 | 0.53 | 0.521 |
0.90 | 0.83 | 0.80 | 0.89 | 0.86 | 0.89 | 0.88 | ||
0.95 | 0.89 | 0.87 | 0.94 | 0.91 | 0.93 | 0.92 | ||
0.99 | 0.96 | 0.95 | 0.99 | 0.97 | 0.99 | 0.99 | ||
25 | 0.50 | 0.48 | 0.47 | 0.49 | 0.48 | 0.53 | 0.53 | |
0.90 | 0.88 | 0.85 | 0.90 | 0.89 | 0.91 | 0.90 | ||
0.95 | 0.93 | 0.91 | 0.95 | 0.93 | 0.95 | 0.94 | ||
0.99 | 0.97 | 0.98 | 0.99 | 0.99 | 1.00 | 0.99 | ||
50 | 0.50 | 0.48 | 0.47 | 0.51 | 0.50 | 0.54 | 0.55 | |
0.90 | 0.89 | 0.87 | 0.91 | 0.88 | 0.90 | 0.91 | ||
0.95 | 0.95 | 0.94 | 0.96 | 0.94 | 0.96 | 0.95 | ||
0.99 | 0.98 | 1.00 | 1.00 | 0.99 | 0.99 | 0.99 | ||
150 | 35 | 0.50 | 0.49 | 0.47 | 0.51 | 0.50 | 0.57 | 0.56 |
0.90 | 0.88 | 0.86 | 0.89 | 0.87 | 0.90 | 0.88 | ||
0.95 | 0.93 | 0.91 | 0.94 | 0.92 | 0.95 | 0.95 | ||
0.99 | 0.98 | 0.99 | 0.99 | 0.99 | 0.99 | 1.00 | ||
50 | 0.50 | 0.50 | 0.49 | 0.55 | 0.54 | 0.58 | 0.57 | |
0.90 | 0.88 | 0.86 | 0.90 | 0.89 | 0.91 | 0.91 | ||
0.95 | 0.94 | 0.93 | 0.95 | 0.94 | 0.95 | 0.95 | ||
0.99 | 0.97 | 0.95 | 0.99 | 0.98 | 0.99 | 1.00 | ||
75 | 0.50 | 0.53 | 0.51 | 0.58 | 0.56 | 0.58 | 0.58 | |
0.90 | 0.89 | 0.88 | 0.91 | 0.90 | 0.93 | 0.93 | ||
0.95 | 0.94 | 0.93 | 0.95 | 0.94 | 0.96 | 0.95 | ||
0.99 | 0.99 | 0.99 | 0.99 | 0.99 | 1.00 | 1.00 |
Average size comparison between MIC and log-likelihood methods
( | ||||||||
---|---|---|---|---|---|---|---|---|
(2.5, 1.5, 0.69, 0.69) | (3, 2, 1.19, 1.19) | (4, 3, 2.19, 2.19) | ||||||
loglik | MIC | loglik | MIC | loglik | MIC | |||
50 | 10 | 0.50 | 7.556 | 7.342 | 3.390 | 3.074 | 2.936 | 2.810 |
0.90 | 9.554 | 9.398 | 3.990 | 3.320 | 3.246 | 3.134 | ||
0.95 | 10.692 | 10.606 | 4.430 | 3.990 | 3.906 | 3.758 | ||
0.99 | 13.794 | 12.734 | 5.686 | 4.634 | 4.368 | 4.146 | ||
15 | 0.50 | 6.774 | 6.526 | 2.996 | 2.620 | 2.788 | 2.584 | |
0.90 | 8.436 | 8.282 | 3.450 | 2.838 | 2.882 | 2.674 | ||
0.95 | 9.462 | 9.364 | 3.902 | 3.604 | 3.566 | 3.350 | ||
0.99 | 11.666 | 11.212 | 4.922 | 4.310 | 3.950 | 3.742 | ||
25 | 0.05 | 5.084 | 4.514 | 2.994 | 2.836 | 2.276 | 2.136 | |
0.90 | 6.852 | 6.338 | 3.422 | 3.392 | 2.366 | 2.242 | ||
0.95 | 7.930 | 7.340 | 3.874 | 3.600 | 2.442 | 2.328 | ||
0.99 | 8.828 | 8.112 | 4.834 | 4.772 | 2.968 | 2.726 | ||
100 | 15 | 0.50 | 6.878 | 6.100 | 2.828 | 2.628 | 2.546 | 2.460 |
0.90 | 8.488 | 8.282 | 3.668 | 3.466 | 2.604 | 2.564 | ||
0.95 | 9.414 | 8.852 | 4.144 | 3.928 | 2.694 | 2.250 | ||
0.99 | 11.960 | 11.640 | 5.052 | 4.436 | 3.386 | 2.918 | ||
25 | 0.50 | 4.986 | 4.810 | 2.392 | 2.096 | 2.230 | 2.166 | |
0.90 | 6.414 | 6.372 | 2.666 | 2.366 | 2.398 | 2.208 | ||
0.95 | 7.250 | 6.814 | 3.120 | 2.840 | 2.516 | 2.268 | ||
0.99 | 9.576 | 8.802 | 3.954 | 3.604 | 2.846 | 2.738 | ||
50 | 0.50 | 4.202 | 4.166 | 2.288 | 2.134 | 2.126 | 2.070 | |
0.90 | 5.616 | 5.310 | 2.538 | 2.340 | 2.354 | 2.214 | ||
0.95 | 6.540 | 6.108 | 3.032 | 2.894 | 2.412 | 2.360 | ||
0.99 | 7.882 | 7.312 | 3.800 | 3.674 | 2.614 | 2.592 | ||
150 | 35 | 0.50 | 3.506 | 2.736 | 2.570 | 2.192 | 2.440 | 2.210 |
0.90 | 4.884 | 4.064 | 2.898 | 2.728 | 2.574 | 2.340 | ||
0.95 | 5.750 | 4.888 | 3.034 | 2.812 | 2.826 | 2.580 | ||
0.99 | 7.972 | 6.950 | 3.808 | 3.596 | 3.468 | 2.874 | ||
50 | 0.50 | 2.524 | 2.288 | 2.268 | 2.146 | 2.218 | 2.122 | |
0.90 | 3.910 | 3.618 | 2.746 | 2.522 | 2.240 | 2.194 | ||
0.95 | 4.716 | 4.480 | 2.842 | 2.766 | 2.774 | 2.476 | ||
0.99 | 6.880 | 6.562 | 3.584 | 3.428 | 2.946 | 2.712 | ||
75 | 0.50 | 2.836 | 2.164 | 2.212 | 2.046 | 2.134 | 1.898 | |
0.90 | 4.162 | 3.710 | 2.494 | 2.254 | 2.358 | 2.234 | ||
0.95 | 4.744 | 4.580 | 2.804 | 2.528 | 2.592 | 2.392 | ||
0.99 | 5.514 | 5.102 | 3.384 | 3.136 | 2.878 | 2.650 |
The consistency of change location estimator
Bias( | MSE( | |||||||
---|---|---|---|---|---|---|---|---|
MIC | BIC | MIC | BIC | MIC | BIC | |||
1 | 50 | 12 | 0.851 | 0.854 | 0.205 | 0.182 | 0.205 | 0.182 |
25 | 0.870 | 0.860 | 0.187 | 0.194 | 0.187 | 0.194 | ||
100 | 25 | 0.883 | 0.879 | 0.196 | 0.199 | 0.196 | 0.199 | |
50 | 0.887 | 0.886 | 0.206 | 0.205 | 0.206 | 0.205 | ||
150 | 37 | 0.914 | 0.887 | 0.201 | 0.185 | 0.201 | 0.185 | |
75 | 0.910 | 0.909 | 0.173 | 0.173 | 0.173 | 0.173 | ||
200 | 50 | 0.895 | 0.910 | 0.191 | 0.199 | 0.191 | 0.199 | |
100 | 0.913 | 0.913 | 0.210 | 0.210 | 0.210 | 0.210 | ||
300 | 75 | 0.892 | 0.904 | 0.169 | 0.196 | 0.169 | 0.196 | |
150 | 0.910 | 0.910 | 0.223 | 0.223 | 0.223 | 0.223 | ||
2 | 50 | 12 | 0.922 | 0.923 | 0.347 | 0.320 | 0.489 | 0.458 |
25 | 0.940 | 0.930 | 0.327 | 0.334 | 0.467 | 0.474 | ||
100 | 25 | 0.956 | 0.952 | 0.342 | 0.345 | 0.488 | 0.491 | |
50 | 0.957 | 0.955 | 0.346 | 0.343 | 0.486 | 0.481 | ||
150 | 37 | 0.972 | 0.956 | 0.317 | 0.323 | 0.433 | 0.461 | |
75 | 0.969 | 0.969 | 0.291 | 0.293 | 0.409 | 0.413 | ||
200 | 50 | 0.966 | 0.962 | 0.333 | 0.303 | 0.475 | 0.407 | |
100 | 0.961 | 0.961 | 0.306 | 0.306 | 0.402 | 0.402 | ||
300 | 75 | 0.964 | 0.966 | 0.303 | 0.320 | 0.457 | 0.444 | |
150 | 0.975 | 0.975 | 0.353 | 0.353 | 0.483 | 0.483 | ||
3 | 50 | 12 | 0.962 | 0.965 | 0.467 | 0.446 | 0.849 | 0.836 |
25 | 0.966 | 0.961 | 0.405 | 0.427 | 0.701 | 0.753 | ||
100 | 25 | 0.973 | 0.976 | 0.393 | 0.417 | 0.641 | 0.707 | |
50 | 0.979 | 0.979 | 0.412 | 0.415 | 0.684 | 0.697 | ||
150 | 37 | 0.985 | 0.981 | 0.356 | 0.398 | 0.550 | 0.686 | |
75 | 0.988 | 0.988 | 0.348 | 0.350 | 0.580 | 0.584 | ||
200 | 50 | 0.988 | 0.989 | 0.399 | 0.384 | 0.673 | 0.650 | |
100 | 0.987 | 0.987 | 0.384 | 0.384 | 0.636 | 0.636 | ||
300 | 75 | 0.985 | 0.981 | 0.376 | 0.365 | 0.646 | 0.579 | |
150 | 0.989 | 0.989 | 0.395 | 0.395 | 0.609 | 0.609 |