
Autoregressive moving average (ARMA) modeling of time series has been successfully applied in different fields such as economics, finance, and engineering. Bayesian analysis of ARMA models is difficult due to the nonlinearity in the coefficients of moving average (MA) part, which complicates the likelihood function and makes the posterior density analytically intractable (Amin, 2019a). Accordingly, the exact posterior analysis of ARMA models requires the use of numerical integration, which is computationally expensive. In order to overcome this problem, three well-known analytic approximations were proposed by Newbold (1973), Zellner and Reynolds (1978), and Broemeling and Shaarawy (1984, 1988) to approximate the posterior density to be an analytically tractable closed-form distribution.
Newbold (1973) proposed the first analytic approximation by expanding unobserved errors as a linear function in the model coefficients using the first order Taylor’s expansion. In addition, Zellner and Reynolds (1978) proposed the second analytic approximation, denoted by Zellner-Reynolds, by expanding the errors sum of squares, rather than the errors, of ARMA model as a quadratic function in the model cofficients using the second order Taylor’s expansion. Moreover, Broemeling and Shaarawy (1984, 1988) proposed the third analytic approximation, denoted by Broemeling-Shaarawy, by simply replacing unobserved lagged errors with their estimates to linearize the errors as a function in the model coefficients.
Because of the simplicity of Broemeling-Shaarawy analytic approximation compared to other proposed approximations, it has been widely extended to the Bayesian analysis of many time series models such as seasonal ARMA models (Amin, 2009; Ismail and Amin, 2014), multivariate MA models (Shaarawy and Ali, 2012), double seasonal MA models (Amin, 2017b), and double seasonal ARMA models (Amin, 2017a, 2018). However, few work have been introduced to evaluate and compare the accuracy of these three analytic approximations. Ismail (1994) used a small scale of simulation study to investigate the accuracy of the three approximations in the case of MA model of order one, and Ali (1998) and Soliman (1999) extended his work to the MA model of order two and to ARMA and seasonal ARMA of order one, respectively. Therefore, there is a need in the Bayesian time series analysis to study how these analytic approximations are mathematically related and to comprehensively evaluate their accuracy, aiming to help researchers to make a trade-off between the simplicity and accuracy of these approximations. As a motivation to this work, the outcome of these analytic approximations for complicated time series models can be used as inputs to advanced methods such as Markov chain Monte Carlo (MCMC) methods (Amin and Ismail, 2015; Amin, 2020, 2022). For example, in the case of seasonal ARMA models where the posterior density is nonlinear function in the model coefficients and thus analytically intractable. So, these analytic approximations can be used to approximate the posterior density, which in turn is used to derive the full conditional posteriors as a main requirement to apply MCMC methods to introduce the Bayesian estimation and prediction for the underlying seasonal ARMA models.
In order to fill this gap, we first review these three analytic approximations and then we use the Kullback-Leibler (KL) divergence (Kullback and Leibler, 1951) to study mathematically the relation between them and to measure the distance between their derived approximate posteriors for ARMA models. In addition, we evaluate the impact of the approximate posteriors distance in the Bayesian estimates of the model coefficients and their precision by generating a large number of Monte Carlo simulations from the approximate posteriors. Therefore, we can summarize our work in this paper as follows. First, we derive the KL divergence between Newbold, Zellner-Reynolds, and Broemeling-Shaarawy approximate posteriors and compute its calibration to study mathematically the relation between them and to measure their distances. Second, we use a large number of Monte Carlo simulations from the approximate posteriors to evaluate the impact of the posteriors distance in the Bayesian estimates of mean and precision of the model coefficients. Finally, we use real-world time series datasets to illustrate the use of the KL divergence to measure the distance between the approximate posteriors and show the impact of this distance in the Bayesian estimates.
The remainder of this paper is organized as follows. In Section 2 we present the background of the ARMA models and related Bayesian concepts, and in Section 3 we summarize the analytic approximations. In Section 4 we derive the KL divergence between the approximate posteriors and its calibration. In Section 5 we present the simulation study and real-world time series datasets to measure the distance between the approximate posteriors and evaluate the impact of this distance in the Bayesian estimates. Finally, we give the conclusions in Section 6.
Time series {
where
where
For ARMA model with normally distributed errors, the likelihood function is given by
where
where
In case of little or no information is available about the model parameters, Jeffreys’ prior can be employed, and it is given as
Multiplying the likelihood function (
For Jeffreys’ prior, the joint posterior of
It is worth observing that the unknown lagged errors are part of the design matrices
Newbold (1973) expanded the errors as a linear function in the model coefficients
where
and
Accordingly, the Newbold approximate errors sum of squares
where
and the Newbold approximate joint posterior of
Using some mathematical manipulations, we can prove that the marginal approximate posterior of the model coefficients
where
Zellner and Reynolds (1978) expanded the errors sum of squares
where
Accordingly, the Zellner-Reynolds approximate likelihood function is given by
and by employing the natural conjugate prior we can obtain the approximate joint posterior of
With some manipulations we can simply derive the marginal approximate posterior of the model coefficients
where
Broemeling and Shaarawy (1984, 1988) approximated the unobserved errors by simply replacing the unobserved lagged errors with their NLSE’s to be a linear function in the model coefficients as follows:
where
and the approximate likelihood function is given by
and the Broemeling-Shaarawy approximate joint posterior of
Using some manipulations, we can derive the marginal approximate posterior of the model coefficients
where
The Kullback-Leibler (KL) divergence (Kullback and Leibler, 1951) can be used to measure the distance between any two of the approximate posteriors of the ARMA model coefficients introduced in Section 3. To explain the idea, suppose we have two multivariate t posteriors for
This KL divergence is not symmetric and always non-negative, and it equals to zero when the two posteriors are the same. This implies that the smaller its value, the closer are the two posteriors. We can simplify the KL divergence (
where CH[
This derived approximate KL divergence is not symmetric; however, we can compute the symmetric distance KL* [
where
Now, we can use this derived approximate KL divergence in
Using the location vectors and dispersion matrices given in
which indicates that the distance between the Newbold and Zellner-Reynolds approximate posteriors depends mainly on the relation between the two matrices
which means (1/2)
which depends on the difference between the two approximate posteriors means and the ratio of the two matrices
It has to be noted that possible values of the approximate KL divergence are non-negative with no maximum limit, however, we can calibrate the values to be in the interval (0.5,1.0) to be able to decide about the distance between the approximate posteriors. When the calibration value closes to 0.5, it implies the two approximate posteriors are almost the same; and when its value closes to 1.0, the two approximate posteriors are strongly different. Suppose KL[
All the derived approximate KL divergences of the approximate posteriors and their calibration will be used in the next section to measure the distance between these posteriors and the impact of this distance in the Bayesian estimates using simulated and real-world time series datasets.
We have two parts of work in this section. First, we present a simulation study to evaluate the KL divergence (and its calibration) between the considered approximate posteriors of the ARMA model coefficients and evaluate the impact of posteriors divergence in the Bayesian estimates. Second, we present two applications of our work to real world time series datasets.
In this simulation study, we have two objectives: (1) evaluating the KL divergence (and its calibration) between the considered approximate posteriors of the ARMA model coefficients, and (2) evaluating the impact of this posteriors divergence in the Bayesian estimates for several simulated time series data with different sample sizes, different ARMA model orders, and different values of the ARMA model coefficients. In the following we discuss the simulation study design and results for each objective.
In order to evaluate the KL divergence (and its calibration) between the considered three approximate posteriors of the ARMA model coefficients, we generate 1,000 time series of size
From these simulation results, we can observe some important remarks.
First, the KL divergence between the Newbold and Zellner-Reynolds approximate posteriors is very small especially for simple models and large sample size, for example their KL calibration values are between 0.57 and 0.71 in most of the cases for
Second, when the sample size increases the approximate posteriors divergences decreases, since more information are provided by the observed time series enabling the analytic approximations to reduce the uncertainty about the model parameters. For example, in case of ARMA(1, 1) with (
Third, the more complicated ARMA model the larger KL divergence between the approximate posteriors. For example, in case of
In order to evaluate the distance impact of the Newbold, Zellner-Reynolds and Broemeling-Shaarawy approximate posteriors in the Bayesian estimates of the ARMA model coefficients
First, we generate 1,000 time series of size
Second, for each time series, we generate 1,000 values of
Third, for each Monte Carlo simulation chain of
Fourth, we compute the mean absolute percentage errors (MAPE) to compare the Bayesian estimates of
Fifth, we evaluate the variability of
Results of the Bayesian estimates of the MA(1) coefficients are presented in Table 6 and their comparison criteria are presented in Table 7. From these results we can observe that the Bayesian estimates of the coefficients mean obtained from all the approximate posteriors are almost the same, however, the standard deviation estimates obtained from the Newbold and Zellner-Reynolds approximate posteriors are strongly different from those obtained form the Broemeling-Shaarawy approximate posterior. For example, when
The results of the comparison criteria show that the Newbold and Zellner-Reynolds approximate posteriors have almost the same MAPE values that are relatively less than those of the Broemeling- Shaarawy posterior. In addition, the ratios of standard deviations show that the standard deviation estimates obtained from the Broemeling-Shaarawy approximate posterior are relatively very large compared to those obtained from the Newbold and Zellner-Reynolds posteriors, which are very close. We get the same conclusions from the results of MA(2) model presented in Tables 8 and 9 and ARMA(1, 1) model presented in Tables 10 and 11 with some additional general remarks. First, when the sample size is getting larger, the Bayesian estimates become more accurate, and hence the MAPE values and the ratios of standard deviation estimates are highly reduced. Second, the larger coefficients values used to generate the time series the more accurate estimates obtained from the approximate posteriors. In general, the simulation results confirm that the Newbold and Zellner-Reynolds approximate posteriors are strongly different from the Broemeling-Shaarawy approximate posterior, and the impact of that can be observed in the Bayesian estimates of the coefficients standard deviation. Thus, the Newbold and Zellner-Reynolds approximations reduce the posterior estimate of the coefficients standard deviation by at least 17% in the case of simple models to 80% in the case of complicated models compared to the Broemeling-Shaarawy approximation, which is strongly reflected in the 95% credible intervals of the coefficients.
We introduce two real-world time series to demonstrate how the KL divergence can be used in reality to measure the distance between the derived approximate posteriors and to show the impact of this distance in the Bayesian estimates of the model coefficients. These two real-world time series datasets represent viscosity and concentration outputs from two different chemical processes (Box
It can observe from Figure 1 that these two time series are not stationary, so we use the first difference to stationarize them; and we then identify the best suitable order of the ARMA model for the differenced time series datasets by using the corrected Akaike’s information criterion (AIC
In this paper we first reviewed the Newbold, Zellner-Reynolds, and Broemeling-Shaarawy analytic approximations proposed in the literature to approximate the posterior density of ARMA models to be analytically tractable. We then derived the KL divergence between these approximate posteriors and compute its calibration to measure the distance between these posteriors. We used a large number of Monte Carlo simulations from the approximate posteriors to evaluate the impact of the posteriors distance in the Bayesian estimates of mean and precision of the model coefficients. Simulation results confirmed that the Newbold and Zellner-Reynolds approximate posteriors are very close to each other, and they are strongly different from the Broemeling-Shaarawy approximate posterior. In particular, the results demonstrated that the Newbold and Zellner-Reynolds approximations show higher precision of the model coefficients compared to the Broemeling-Shaarawy approximation, which is reflected in the coefficients credible intervals. We applied our work to real-world time series datasets and their results are consistent with those of the simulation study.
Results of KL divergence and its calibration (KLC) for MA(1)
Sample size ( | |||||||
---|---|---|---|---|---|---|---|
50 | 100 | 150 | 200 | 250 | 300 | ||
0.4 | KL[N,ZR]* | 0.067 | 0.030 | 0.019 | 0.014 | 0.011 | 0.009 |
KLC[N,ZR] | 0.661 | 0.614 | 0.593 | 0.580 | 0.571 | 0.565 | |
KL[N,BS] | 0.373 | 0.136 | 0.123 | 0.111 | 0.105 | 0.102 | |
KLC[N,BS] | 0.725 | 0.703 | 0.695 | 0.689 | 0.685 | 0.683 | |
KL[ZR,BS] | 0.465 | 0.139 | 0.127 | 0.116 | 0.108 | 0.105 | |
KLC[ZR,BS] | 0.738 | 0.708 | 0.699 | 0.692 | 0.686 | 0.684 | |
0.7 | KL[N,ZR] | 0.098 | 0.151 | 0.027 | 0.021 | 0.016 | 0.013 |
KLC[N,ZR] | 0.679 | 0.628 | 0.604 | 0.592 | 0.582 | 0.574 | |
KL[N,BS] | 1.244 | 0.703 | 0.546 | 0.554 | 0.545 | 0.538 | |
KLC[N,BS] | 0.864 | 0.856 | 0.852 | 0.853 | 0.851 | 0.850 | |
KL[ZR,BS] | 1.406 | 0.601 | 0.545 | 0.545 | 0.532 | 0.531 | |
KLC[ZR,BS] | 0.860 | 0.852 | 0.848 | 0.850 | 0.848 | 0.848 |
*N, ZR and BS refer to Newbold, Zellner-Reynolds and Broemeling-Shaarawy approximate posteriors, respectively.
Results of KL divergence and its calibration (KLC) for MA(2)
( | Sample size ( | ||||||
---|---|---|---|---|---|---|---|
50 | 100 | 150 | 200 | 250 | 300 | ||
(0.4, 0.3) | KL[N,ZR] | 0.188 | 0.083 | 0.053 | 0.039 | 0.031 | 0.025 |
KLC[N,ZR] | 0.763 | 0.688 | 0.653 | 0.632 | 0.618 | 0.607 | |
KL[N,BS] | 0.618 | 0.274 | 0.236 | 0.217 | 0.211 | 0.205 | |
KLC[N,BS] | 0.833 | 0.798 | 0.784 | 0.775 | 0.772 | 0.769 | |
KL[ZR,BS] | 0.726 | 0.301 | 0.253 | 0.229 | 0.221 | 0.213 | |
KLC[ZR,BS] | 0.851 | 0.808 | 0.790 | 0.780 | 0.776 | 0.773 | |
(0.3, 0.6) | KL[N,ZR] | 0.482 | 0.116 | 0.068 | 0.049 | 0.039 | 0.032 |
KLC[N,ZR] | 0.789 | 0.708 | 0.669 | 0.645 | 0.631 | 0.619 | |
KL[N,BS] | 2.176 | 0.805 | 0.693 | 0.641 | 0.625 | 0.610 | |
KLC[N,BS] | 0.910 | 0.895 | 0.891 | 0.886 | 0.883 | 0.882 | |
KL[ZR,BS] | 2.230 | 0.822 | 0.698 | 0.645 | 0.628 | 0.611 | |
KLC[ZR,BS] | 0.918 | 0.898 | 0.893 | 0.887 | 0.884 | 0.883 |
Results of KL divergence and its calibration (KLC) for ARMA(1,1)
( | Sample size ( | ||||||
---|---|---|---|---|---|---|---|
50 | 100 | 150 | 200 | 250 | 300 | ||
(0.5, 0.3) | KL[N,ZR] | 0.145 | 0.073 | 0.044 | 0.033 | 0.026 | 0.022 |
KLC[N,ZR] | 0.738 | 0.673 | 0.640 | 0.623 | 0.610 | 0.600 | |
KL[N,BS] | 0.625 | 0.204 | 0.164 | 0.143 | 0.131 | 0.128 | |
KLC[N,BS] | 0.807 | 0.761 | 0.742 | 0.732 | 0.726 | 0.723 | |
KL[ZR,BS] | 0.795 | 0.231 | 0.180 | 0.156 | 0.142 | 0.137 | |
KLC[ZR,BS] | 0.825 | 0.775 | 0.752 | 0.740 | 0.732 | 0.729 | |
(0.4, 0.6) | KL[N,ZR] | 0.184 | 0.074 | 0.046 | 0.035 | 0.027 | 0.022 |
KLC[N,ZR] | 0.743 | 0.676 | 0.642 | 0.624 | 0.611 | 0.601 | |
KL[N,BS] | 1.526 | 0.748 | 0.576 | 0.550 | 0.544 | 0.540 | |
KLC[N,BS] | 0.906 | 0.893 | 0.885 | 0.882 | 0.881 | 0.882 | |
KL[ZR,BS] | 1.719 | 0.730 | 0.598 | 0.561 | 0.550 | 0.547 | |
KLC[ZR,BS] | 0.911 | 0.894 | 0.885 | 0.882 | 0.881 | 0.882 |
Results of KL divergence and its calibration (KLC) for ARMA(1,2)
( | Sample size ( | ||||||
---|---|---|---|---|---|---|---|
50 | 100 | 150 | 200 | 250 | 300 | ||
(0.4, 0.5, 0.4) | KL[N,ZR] | 0.440 | 0.198 | 0.107 | 0.080 | 0.062 | 0.048 |
KLC[N,ZR] | 0.841 | 0.755 | 0.710 | 0.683 | 0.662 | 0.647 | |
KL[N,BS] | 1.991 | 0.873 | 0.722 | 0.666 | 0.649 | 0.628 | |
KLC[N,BS] | 0.937 | 0.923 | 0.917 | 0.912 | 0.910 | 0.909 | |
KL[ZR,BS] | 2.236 | 0.930 | 0.758 | 0.695 | 0.672 | 0.648 | |
KLC[ZR,BS] | 0.948 | 0.927 | 0.921 | 0.915 | 0.913 | 0.911 | |
(0.7, 0.3, 0.4) | KL[N,ZR] | 0.359 | 0.145 | 0.091 | 0.066 | 0.050 | 0.041 |
KLC[N,ZR] | 0.828 | 0.742 | 0.698 | 0.672 | 0.652 | 0.638 | |
KL[N,BS] | 1.334 | 0.661 | 0.565 | 0.524 | 0.502 | 0.484 | |
KLC[N,BS] | 0.927 | 0.902 | 0.894 | 0.886 | 0.883 | 0.879 | |
KL[ZR,BS] | 1.467 | 0.701 | 0.591 | 0.545 | 0.517 | 0.498 | |
KLC[ZR,BS] | 0.937 | 0.909 | 0.898 | 0.890 | 0.886 | 0.882 |
Results of KL divergence and its calibration (KLC) for ARMA(2,2)
( | Sample size ( | ||||||
---|---|---|---|---|---|---|---|
50 | 100 | 150 | 200 | 250 | 300 | ||
(0.4, 0.3, 0.5, 0.4) | KL[N,ZR] | 2.537 | 0.370 | 0.239 | 0.207 | 0.136 | 0.109 |
KLC[N,ZR] | 0.902 | 0.817 | 0.776 | 0.749 | 0.724 | 0.706 | |
KL[N,BS] | 4.038 | 1.532 | 1.148 | 0.913 | 0.864 | 0.825 | |
KLC[N,BS] | 0.957 | 0.928 | 0.925 | 0.921 | 0.920 | 0.919 | |
KL[ZR,BS] | 6.679 | 1.696 | 1.270 | 1.030 | 0.915 | 0.876 | |
KLC[ZR,BS] | 0.970 | 0.943 | 0.936 | 0.934 | 0.928 | 0.926 | |
(1.3, 0.6, 0.3, 0.6) | KL[N,ZR] | 0.595 | 0.257 | 0.152 | 0.111 | 0.085 | 0.070 |
KLC[N,ZR] | 0.887 | 0.798 | 0.747 | 0.715 | 0.692 | 0.676 | |
KL[N,BS] | 4.310 | 1.691 | 1.496 | 1.425 | 1.334 | 1.297 | |
KLC[N,BS] | 0.977 | 0.975 | 0.973 | 0.972 | 0.970 | 0.969 | |
KL[ZR,BS] | 4.652 | 1.701 | 1.527 | 1.457 | 1.359 | 1.320 | |
KLC[ZR,BS] | 0.981 | 0.977 | 0.975 | 0.973 | 0.971 | 0.969 |
Bayesian estimates of the coefficients of MA(1)
0.4 | 50 | 0.413 | 0.016 | 0.108 | 0.715 | 0.413 | 0.018 | 0.108 | 0.715 | 0.410 | 0.020 | 0.107 | 0.708 |
100 | 0.406 | 0.008 | 0.210 | 0.607 | 0.406 | 0.009 | 0.210 | 0.607 | 0.402 | 0.010 | 0.201 | 0.583 | |
150 | 0.405 | 0.006 | 0.242 | 0.562 | 0.405 | 0.006 | 0.242 | 0.562 | 0.403 | 0.007 | 0.236 | 0.567 | |
200 | 0.403 | 0.004 | 0.266 | 0.533 | 0.403 | 0.004 | 0.266 | 0.533 | 0.402 | 0.005 | 0.254 | 0.540 | |
250 | 0.402 | 0.003 | 0.283 | 0.520 | 0.402 | 0.003 | 0.283 | 0.520 | 0.401 | 0.004 | 0.272 | 0.530 | |
300 | 0.403 | 0.003 | 0.301 | 0.506 | 0.403 | 0.003 | 0.301 | 0.506 | 0.402 | 0.003 | 0.291 | 0.517 | |
0.7 | 50 | 0.723 | 0.010 | 0.483 | 1.000 | 0.723 | 0.011 | 0.483 | 1.000 | 0.715 | 0.020 | 0.405 | 1.007 |
100 | 0.710 | 0.005 | 0.546 | 0.867 | 0.710 | 0.006 | 0.546 | 0.867 | 0.704 | 0.010 | 0.503 | 0.883 | |
150 | 0.707 | 0.003 | 0.576 | 0.831 | 0.707 | 0.004 | 0.576 | 0.831 | 0.704 | 0.007 | 0.536 | 0.862 | |
200 | 0.705 | 0.003 | 0.596 | 0.809 | 0.705 | 0.003 | 0.596 | 0.809 | 0.703 | 0.005 | 0.552 | 0.840 | |
250 | 0.704 | 0.002 | 0.614 | 0.795 | 0.704 | 0.002 | 0.614 | 0.795 | 0.702 | 0.004 | 0.574 | 0.828 | |
300 | 0.704 | 0.002 | 0.628 | 0.786 | 0.704 | 0.002 | 0.628 | 0.786 | 0.702 | 0.003 | 0.591 | 0.814 |
Comparison criteria between estimates from different posteriors for MA(1).
MAPE | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
0.4 | 50 | 58.331 | 58.331 | 60.645 | 1.132 | 0.630 | 2.348 | 1.399 | 1.013 | 2.153 |
100 | 39.391 | 39.391 | 40.614 | 1.066 | 0.708 | 1.735 | 1.223 | 1.042 | 1.570 | |
150 | 31.481 | 31.481 | 33.127 | 1.040 | 0.743 | 1.523 | 1.212 | 1.061 | 1.465 | |
200 | 27.033 | 27.033 | 29.372 | 1.030 | 0.772 | 1.421 | 1.205 | 1.074 | 1.395 | |
250 | 23.999 | 23.999 | 26.294 | 1.025 | 0.798 | 1.349 | 1.202 | 1.087 | 1.368 | |
300 | 21.747 | 21.747 | 23.791 | 1.022 | 0.809 | 1.322 | 1.200 | 1.099 | 1.342 | |
0.7 | 50 | 28.388 | 28.388 | 34.865 | 1.276 | 0.549 | 3.199 | 3.520 | 1.261 | 19.866 |
100 | 18.299 | 18.299 | 23.246 | 1.120 | 0.651 | 2.116 | 2.234 | 1.412 | 4.095 | |
150 | 14.496 | 14.496 | 18.943 | 1.066 | 0.686 | 1.752 | 2.101 | 1.485 | 3.187 | |
200 | 12.217 | 12.217 | 16.757 | 1.055 | 0.723 | 1.653 | 2.042 | 1.541 | 2.824 | |
250 | 10.532 | 10.532 | 14.989 | 1.047 | 0.742 | 1.591 | 2.025 | 1.601 | 2.702 | |
300 | 9.541 | 9.541 | 13.526 | 1.035 | 0.754 | 1.489 | 2.014 | 1.645 | 2.598 |
Bayesian estimates of the coefficients of MA(2)
50 | 0.310 | 0.012 | 0.049 | 0.577 | 0.310 | 0.017 | 0.049 | 0.577 | 0.310 | 0.021 | 0.002 | 0.622 | |
100 | 0.303 | 0.006 | 0.139 | 0.471 | 0.303 | 0.007 | 0.139 | 0.471 | 0.301 | 0.010 | 0.099 | 0.486 | |
150 | 0.303 | 0.004 | 0.163 | 0.438 | 0.303 | 0.005 | 0.163 | 0.438 | 0.302 | 0.007 | 0.139 | 0.459 | |
200 | 0.301 | 0.003 | 0.184 | 0.419 | 0.301 | 0.003 | 0.184 | 0.419 | 0.302 | 0.005 | 0.154 | 0.438 | |
250 | 0.301 | 0.003 | 0.193 | 0.404 | 0.301 | 0.003 | 0.193 | 0.404 | 0.301 | 0.004 | 0.173 | 0.427 | |
300 | 0.301 | 0.002 | 0.207 | 0.395 | 0.301 | 0.002 | 0.207 | 0.395 | 0.302 | 0.003 | 0.189 | 0.417 | |
50 | 0.641 | 0.012 | 0.330 | 1.000 | 0.641 | 0.015 | 0.330 | 1.000 | 0.621 | 0.021 | 0.298 | 0.956 | |
100 | 0.611 | 0.006 | 0.419 | 0.808 | 0.611 | 0.008 | 0.419 | 0.808 | 0.603 | 0.010 | 0.387 | 0.823 | |
150 | 0.606 | 0.004 | 0.471 | 0.745 | 0.606 | 0.005 | 0.471 | 0.745 | 0.601 | 0.007 | 0.429 | 0.772 | |
200 | 0.603 | 0.003 | 0.481 | 0.725 | 0.603 | 0.003 | 0.481 | 0.725 | 0.600 | 0.005 | 0.459 | 0.744 | |
250 | 0.602 | 0.003 | 0.500 | 0.703 | 0.602 | 0.003 | 0.500 | 0.703 | 0.599 | 0.004 | 0.480 | 0.721 | |
300 | 0.601 | 0.002 | 0.514 | 0.690 | 0.601 | 0.002 | 0.514 | 0.690 | 0.598 | 0.003 | 0.491 | 0.708 |
Comparison criteria between estimates from different posteriors for MA(2)
MAPE | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
50 | 71.962 | 71.962 | 80.500 | 1.730 | 0.284 | 3.258 | 2.820 | 1.084 | 17.998 | |
100 | 45.565 | 45.565 | 53.860 | 1.094 | 0.799 | 1.662 | 1.733 | 1.180 | 2.824 | |
150 | 36.249 | 36.249 | 43.715 | 1.059 | 0.828 | 1.408 | 1.614 | 1.270 | 2.226 | |
200 | 32.125 | 32.125 | 38.935 | 1.043 | 0.840 | 1.328 | 1.593 | 1.294 | 2.106 | |
250 | 28.718 | 28.718 | 34.831 | 1.037 | 0.857 | 1.300 | 1.584 | 1.327 | 1.978 | |
300 | 25.837 | 25.837 | 31.518 | 1.030 | 0.875 | 1.255 | 1.575 | 1.349 | 1.890 | |
50 | 44.276 | 44.276 | 44.943 | 1.223 | 0.489 | 4.470 | 2.920 | 1.097 | 20.293 | |
100 | 25.730 | 25.730 | 29.330 | 1.179 | 0.704 | 2.301 | 1.754 | 1.198 | 2.874 | |
150 | 19.218 | 19.218 | 23.505 | 1.098 | 0.733 | 1.770 | 1.626 | 1.273 | 2.276 | |
200 | 16.030 | 16.030 | 19.674 | 1.066 | 0.763 | 1.566 | 1.602 | 1.300 | 2.116 | |
250 | 14.045 | 14.045 | 17.280 | 1.049 | 0.783 | 1.483 | 1.591 | 1.332 | 1.998 | |
300 | 12.431 | 12.431 | 15.695 | 1.039 | 0.792 | 1.395 | 1.581 | 1.351 | 1.904 |
Bayesian estimates of the coefficients of ARMA(1, 1)
50 | 0.192 | 0.113 | 0.733 | 0.645 | 0.192 | 0.088 | 0.733 | 0.645 | 0.249 | 0.196 | 0.858 | 0.957 | |
100 | 0.274 | 0.035 | 0.205 | 0.615 | 0.274 | 0.036 | 0.205 | 0.615 | 0.311 | 0.058 | 0.226 | 0.762 | |
150 | 0.291 | 0.023 | 0.071 | 0.584 | 0.291 | 0.032 | 0.071 | 0.584 | 0.316 | 0.039 | 0.116 | 0.711 | |
200 | 0.294 | 0.018 | 0.008 | 0.541 | 0.294 | 0.019 | 0.008 | 0.541 | 0.317 | 0.029 | 0.061 | 0.672 | |
250 | 0.297 | 0.014 | 0.033 | 0.521 | 0.297 | 0.014 | 0.033 | 0.521 | 0.314 | 0.023 | 0.005 | 0.622 | |
300 | 0.299 | 0.012 | 0.059 | 0.513 | 0.299 | 0.012 | 0.059 | 0.513 | 0.317 | 0.019 | 0.046 | 0.601 | |
50 | 0.628 | 0.091 | 1.000 | 0.540 | 0.628 | 0.073 | 1.000 | 0.540 | 0.674 | 0.218 | 1.209 | 0.676 | |
100 | 0.698 | 0.022 | 1.000 | 0.195 | 0.698 | 0.025 | 1.000 | 0.195 | 0.729 | 0.068 | 1.062 | 0.207 | |
150 | 0.705 | 0.014 | 0.924 | 0.370 | 0.705 | 0.020 | 0.924 | 0.370 | 0.725 | 0.046 | 1.011 | 0.338 | |
200 | 0.703 | 0.010 | 0.887 | 0.458 | 0.703 | 0.011 | 0.887 | 0.458 | 0.723 | 0.034 | 0.982 | 0.417 | |
250 | 0.705 | 0.008 | 0.867 | 0.472 | 0.705 | 0.008 | 0.867 | 0.472 | 0.720 | 0.027 | 0.944 | 0.463 | |
300 | 0.704 | 0.007 | 0.864 | 0.509 | 0.704 | 0.007 | 0.864 | 0.509 | 0.720 | 0.023 | 0.931 | 0.480 |
Comparison criteria between estimates from different posterior s for ARMA(1, 1)
MAPE | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
50 | 84.462 | 84.462 | 100.995 | 0.712 | 0.390 | 3.296 | 2.492 | 0.889 | 8.979 | |
100 | 55.174 | 55.174 | 67.819 | 0.968 | 0.595 | 2.942 | 2.081 | 0.931 | 5.954 | |
150 | 42.149 | 42.149 | 52.445 | 2.265 | 0.665 | 2.592 | 1.901 | 0.961 | 4.208 | |
200 | 36.603 | 36.603 | 45.429 | 1.083 | 0.678 | 2.081 | 1.777 | 1.018 | 3.595 | |
250 | 31.576 | 31.576 | 39.450 | 1.037 | 0.700 | 1.774 | 1.748 | 1.050 | 3.272 | |
300 | 29.338 | 29.338 | 36.662 | 1.028 | 0.724 | 1.710 | 1.724 | 1.097 | 3.065 | |
50 | 34.121 | 34.121 | 41.079 | 0.809 | 0.355 | 5.575 | 10.055 | 1.048 | 72.894 | |
100 | 19.429 | 19.429 | 24.556 | 0.911 | 0.445 | 4.614 | 9.783 | 1.200 | 70.867 | |
150 | 14.694 | 14.694 | 19.314 | 0.978 | 0.518 | 3.485 | 7.235 | 1.369 | 21.411 | |
200 | 11.755 | 11.755 | 15.667 | 1.141 | 0.558 | 2.443 | 4.385 | 1.611 | 13.234 | |
250 | 9.975 | 9.975 | 13.448 | 1.056 | 0.596 | 2.086 | 4.189 | 1.670 | 10.889 | |
300 | 9.274 | 9.274 | 12.541 | 1.055 | 0.614 | 1.960 | 4.122 | 1.779 | 10.158 |
Results of KL divergence and its calibration for real-world time series datasets
Dataset | KL[ | KLC[ | KL[ | KLC[ | KL[ZR, BS] | KLC[ZR, BS] |
---|---|---|---|---|---|---|
Series A | 0.0248 | 0.6100 | 2.1950 | 0.9969 | 1.9265 | 0.9947 |
Series B | 0.0273 | 0.6153 | 0.1145 | 0.7262 | 0.1088 | 0.7211 |
Bayesian estimates of the coefficients for real-world time series datasets
Dataset | |||||||||
---|---|---|---|---|---|---|---|---|---|
Series A | 0.2155 | 0.0092 | 0.2155 | 0.0092 | 0.1142 | 0.0138 | 1.0086 | 1.5027 | |
0.8193 | 0.0032 | 0.8193 | 0.0035 | 0.7267 | 0.0190 | 1.1136 | 5.9836 | ||
Series B | 0.4613 | 0.1723 | 0.4613 | 0.1953 | 0.4082 | 0.2375 | 1.1338 | 1.3789 | |
0.3246 | 0.1956 | 0.3246 | 0.2228 | 0.2666 | 0.2431 | 1.1390 | 1.2430 |