NML, Bayes and true distributions: A comment on Karabatsos and Walker (2006)

作者: Peter Grünwald , Daniel J. Navarro

DOI: 10.1016/J.JMP.2008.11.005

关键词:

摘要: Abstract We review the normalized maximum likelihood (NML) criterion for selecting among competing models. NML is generally justified on information-theoretic grounds, via principle of minimum description length (MDL), in a derivation that “does not assume existence true, data-generating distribution”. Since this “agnostic” claim has been source some recent confusion psychological literature, we explain detail what meant by statement. In doing so discuss work presented [Karabatsos, G., & Walker, S. G. (2006). On and Bayesian decision theory. Journal Mathematical Psychology, 50, 517–520], who propose an alternative decision-theoretic characterization NML, which leads them to conclude agnosticity meaningless. KW derivation, one part (the term) arises from placing Dirichlet process prior over possible distributions, other complexity folded into loss function. Whereas original derivations term naturally, KW derivation its mathematical form taken granted explained any further. argue reason, KW characterization incomplete; relatedly, question relevance their main conclusion about does follow.

参考文章(23)
Peter D. Grünwald, In Jae Myung, Mark A. Pitt, Advances in Minimum Description Length: Theory and Applications MIT Press. ,(2005)
Paul E. Meehl, Why Summaries of Research on Psychological Theories are Often Uninterpretable Psychological Reports. ,vol. 66, pp. 195- 244 ,(1990) , 10.2466/PR0.1990.66.1.195
Sonia Petrone, Adrian E. Raftery, A note on the Dirichlet process prior in Bayesian nonparametric inference with partial exchangeability Statistics & Probability Letters. ,vol. 36, pp. 69- 83 ,(1997) , 10.1016/S0167-7152(97)00050-3
Jay I. Myung, Daniel J. Navarro, Mark A. Pitt, Model selection by normalized maximum likelihood Journal of Mathematical Psychology. ,vol. 50, pp. 167- 179 ,(2006) , 10.1016/J.JMP.2005.06.008
Michael W Browne, Cross-validation methods Journal of Mathematical Psychology. ,vol. 44, pp. 108- 132 ,(2000) , 10.1006/JMPS.1999.1279
Hirotogu Akaike, Information Theory and an Extension of the Maximum Likelihood Principle international symposium on information theory. ,vol. 1, pp. 610- 624 ,(1973) , 10.1007/978-1-4612-1694-0_15
Thomas S. Ferguson, A Bayesian Analysis of Some Nonparametric Problems Annals of Statistics. ,vol. 1, pp. 209- 230 ,(1973) , 10.1214/AOS/1176342360
Peter Grünwald, John Langford, Suboptimal behavior of Bayes and MDL in classification under misspecification Machine Learning. ,vol. 66, pp. 119- 149 ,(2007) , 10.1007/S10994-007-0716-7
Michael D. Escobar, Mike West, Bayesian Density Estimation and Inference Using Mixtures Journal of the American Statistical Association. ,vol. 90, pp. 577- 588 ,(1995) , 10.1080/01621459.1995.10476550