You can edit almost every page by Creating an account. Otherwise, see the FAQ.

Frequentist model averaging

From EverybodyWiki Bios & Wiki

Script error: No such module "AfC submission catcheck".


In prediction, considering multiple models, rather than just an individual model, is a way to manage the uncertainty about which of a set of individual models is likely to give the best results. [1] [2] [3]

Model selection methods are a way to choose just one of multiple models and reject the rest. Model averaging methods are a way to combine multiple models in a weighted combination. Model averaging may have advantages over model selection, especially when (a) several of the models are likely to perform similarly well, and choosing just one of them is a fairly arbitrary choice, or (b) one is interested in ensuring that the entire range of reasonable possible predictions is captured, for example in risk management. On the other hand, model selection may have advantages over model averaging in a situation in which one of the models is likely to perform much better than all the other models.

Model averaging can be performed based on either frequentist or Bayesian ideas, or a mixture of the two. There are pros and cons to the two approaches. [4]

This article is about frequentist model averaging (FMA) methods.

Types of Frequentist Model Averaging[edit]

What follows is a list of types of frequentist model averaging taken from the recent academic literature. The list is not exhaustive, and some of the items overlap. The list is ordered from simple, standard and well-used methods to more complex, more unusual methods.

Plug-in model averaging (PMA)[edit]

PMA [5] [6] [7] [8] is perhaps the simplest form of frequentist model averaging. The goal of PMA is to produce more accurate point forecasts, where more accurate is defined as lower predictive mean squared error (PMSE). To derive PMA, an expression for the predictive mean squared error PMSE is written in terms of the model weights and the true values for the model parameters. This expression is then maximised by varying the weights, giving the weights that would maximise PMSE, as a function of the true values for the model parameters. The true values for the model parameters are then replaced by the estimated values for the model parameters (the estimated values are simply ‘plugged in’). PMA has the advantage that the derivation is the simplest and easiest to understand of any of the FMA methods. Some authors have reported good results from PMA, [5] [6] [9] including for probabilistic forecasting. [8]

AIC model averaging (AICMA)[edit]

In AIC model averaging, models fitted using maximum likelihood are combined using weights derived from the Akaike information criterion (AIC) score, known as Akaike weights. [1] [4] [10] [11] [7] The goal of AICMA is to produce more accurate probabilistic forecasts, where more accurate is defined as lower Kullback-Leibler divergence. AICMA has the advantages that AIC is well known and well accepted, and that AICMA is simple. It has the disadvantage that it only applies to models fitted using maximum likelihood estimation. AICMA can also be based on AICc, which is a small sample variant of AIC. There is also a variant of AICMA known as smoothed AICMA.

BIC model averaging (BICMA)[edit]

In BIC model averaging, models fitted using maximum likelihood are combined using weights based on the Bayesian information criterion (BIC) score. [1] [10] [11] [7]. Like AICMA, BICMA has the advantages that BIC is well known and well accepted, and that BICMA is simple. It again has the disadvantage that it only applies to models fitted using maximum likelihood estimation. There is also a variant of BICMA known as smoothed BICMA.

Cross-validation model averaging (CVMA)[edit]

In CVMA, cross-validation is used to determine weights for a given criterion. [11] [6] [7] Many different criteria are possible, such as predictive mean squared error, or predictive mean log-likelihood (PMLL). CVMA can also be applied to models fitted in different ways (i.e., is not restricted to maximum likelihood models). This flexibility of criterion and models is the big advantage of CVMA relative to e.g., AIC and BIC.

Jack-knife model averagin (JMA)[edit]

JMA is a special case of CVMA, in which the cross-validation is performed using a jack-knife. [11] [6] [7]

Focussed model averaging (FMA)[edit]

[4]. [6]

Least Squares model averaging (LSMA)[edit]

LSMA methods are a class of model averaging methods that are designed to minimise the predictive mean squared error (PMSE), by averaging across models fitted using least squares. [5] [9] [7] [8] Note that PMSE is known as the mean square forecast error (MSFE). Example of LSMA methods from this list are MMA, PMA, and also JMA if the criterion used in the JMA is predictive mean squared error (PMSE).

LSMA has also been known as mean squared error model averaging, although using the name least squares model averaging and the abbreviation LSMA avoids a clash with the abbreviation MMA which has been used for Mallows model averaging.

Mallows model averaging (MMA)[edit]

In MMA, models fitted using least squares are combined using weights derived from Mallows Cp. [10] [11] [6] MMA is a type of LSMA.

Bayesian plug-in model averaging (BPMA)[edit]

Bayesian plug-in model averaging is an extension of PMA, in which the weights are calculated as the Bayesian expectation over all possible values for the weights. [8].

WAIC model averaging (WMA)[edit]

In WMA model averaging, models are combined using weights based on the WAIC score instead of the AIC score.

Performance[edit]

In simulation tests, no model averaging method dominates the other methods for all values of the real parameters. Typically each method performs well for a range of real parameter values. Simulation results are given in many places in the literature.

Applications in Economics[edit]

Applications in Atmospheric Science[edit]

FMA methods have been used to estimate the rate of climate change and make predictions of future climate. Since estimates of the rate of climate change may be highly uncertain, for certain climate variables, the best predictions may come from any of the following three methods:

  • assuming that the rate of climate change is zero. This is likely to give the best results when the rate of climate change is so hard to estimate that assuming it is zero gives better predictions. This is typically when the uncertainty around the estimate of the rate is much greater than the rate itself.
  • estimating the rate of climate change using unbiased estimation methods (such as maximum likelihood). This is likely to give the best results when the rate of climate change can be estimated relatively accurately. This is typically when the uncertainty around the estimate of the rate is much less than the rate itself.
  • using FMA to combine a change of zero with an unbiased estimate, giving something in-between. This is likely to give the best results when the rate of climate change is roughly similar to the uncertainty around the estimate of the rate.

Specific examples from the atmospheric science literature include:

  • Estimating the recent trend in local temperatures from local observations using PMA to combine ‘no trend’ and ‘OLS trend’.[5]
  • Estimating the recent trend in regional rainfall indices from observations using JMA, AICMA, BICMA and PMA.[7]
  • Estimating the future change in regional rainfall indices from climate models using AICMA, PMA and BPMA.[8]
  • add new reference

References[edit]

  1. 1.0 1.1 1.2 Burnham, K; Anderson, D (2002). Model Selection and Multimodel Inference. Springer. Search this book on
  2. Claeskens, G; Hjort, N (2008). Model Selection and Model Averaging. CUP. Search this book on
  3. Fletcher, D (2019). Model Averaging. Springer. Search this book on
  4. 4.0 4.1 4.2 Hjort, N.; Claeskens, G. (2003). "Frequentist model average estimators". Journal of the American Statistical Association. 98 (x): 879–899. doi:10.1198/016214503000000828. hdl:10852/10306.
  5. 5.0 5.1 5.2 5.3 Jewson, S.; Penzer, J. (2006). "Estimating Trends in Weather Series: Consequences for Pricing Derivatives". Studies in Nonlinear Dynamics and Econometrics. 10 (3): TBD. doi:10.2202/1558-3708.1386.
  6. 6.0 6.1 6.2 6.3 6.4 6.5 Liu, C. (2014). "Distribution theory of the least squares averaging estimator" (PDF). Journal of Econometrics. 186 (1): 142–159. doi:10.1016/j.jeconom.2014.07.002.
  7. 7.0 7.1 7.2 7.3 7.4 7.5 7.6 Jewson, S.; Dallafior, T.; Comola, F. (2021). "Dealing with Trend Uncertainty in Empirical Estimates of European Rainfall Climate for Insurance Risk Management". Meteorological Applications. TBD (TBD): TBD.
  8. 8.0 8.1 8.2 8.3 8.4 Jewson, S.; Barbato, G.; Mercogliano, P.; Mysiak, J.; Sassi, M. (2021). "Improving the Potential Accuracy and Usability of EURO-CORDEX Estimates of Future Rainfall Climate using Mean Squared Error Model Averaging". Nonlinear Processes in Geophysics. TBD (TBD): TBD. doi:10.21203/rs.3.rs-494689/v1.
  9. 9.0 9.1 Charkhi, A.; Claeskens, G.; Hansen, B. (2016). "Minimum mean squared error model averaging in likelihood models". Statistica Sinica. 26: 809–840. doi:10.5705/ss.202014.0067.
  10. 10.0 10.1 10.2 Hansen, B. (2007). "Least Squares Model Averaging". Econometrica. 75 (4): 1175–1189. doi:10.1111/j.1468-0262.2007.00785.x.
  11. 11.0 11.1 11.2 11.3 11.4 Hansen, B.; Racine, J. (2012). "Jackknife Model Averaging". Journal of Econometrics. 167: 38–46. doi:10.1016/j.jeconom.2011.06.019.


This article "Frequentist model averaging" is from Wikipedia. The list of its authors can be seen in its historical and/or the page Edithistory:Frequentist model averaging. Articles copied from Draft Namespace on Wikipedia could be seen on the Draft Namespace of Wikipedia and not main one.