×

Deep fiducial inference. (English) Zbl 07851194

Summary: Since the mid-2000s, there has been a resurrection of interest in modern modifications of fiducial inference. To date, the main computational tool to extract a generalized fiducial distribution is Markov chain Monte Carlo (MCMC). We propose an alternative way of computing a generalized fiducial distribution that could be used in complex situations. In particular, to overcome the difficulty when the unnormalized fiducial density (needed for MCMC) is intractable, we design a fiducial autoencoder (FAE). The fitted FAE is used to generate generalized fiducial samples of the unknown parameters. To increase accuracy, we then apply an approximate fiducial computation (AFC) algorithm, by rejecting samples that when plugged into a decoder do not replicate the observed data well enough. Our numerical experiments show the effectiveness of our FAE-based inverse solution and the excellent coverage performance of the AFC-corrected FAE solution.
{© 2020 John Wiley & Sons, Ltd.}

MSC:

62-XX Statistics

References:

[1] Bardsley, J. M., Solonen, A., Haario, H., & Laine, M. (2014). Randomize‐then‐optimize: A method for sampling from posterior distributions in nonlinear inverse problems. SIAM Journal on Scientific Computing, 36(4). https://doi.org/10.1137/140964023 · Zbl 1303.65003 · doi:10.1137/140964023
[2] Birnbaum, A (1961). On the foundations of statistical inference: Binary experiments. The Annals of Mathematical Statistics, 414-435. · Zbl 0118.13704
[3] Blum, M. G., Nunes, M. A., Prangle, D., & Sisson, S. A. (2013). A comparative review of dimension reduction methods in approximate bayesian computation. Statistical Science, 28(2), 189-208. · Zbl 1331.62123
[4] Bottou, L. (2010). Large‐scale machine learning with stochastic gradient descent. In Proceedings of COMPSTAT’2010, Springer, pp. 177-186. · Zbl 1436.68293
[5] Chollet, F., & Others. (2015). Keras. https://keras.io
[6] Doersch, C. (2016). Tutorial on variational autoencoders. arXiv preprint arXiv:1606.05908.
[7] Fisher, R. A. (1930). Inverse probability. Mathematical Proceedings of the Cambridge Philosophical Society, 26(4), 528-535. https://doi.org/10.1017/s0305004100016297 · JFM 56.1083.05 · doi:10.1017/s0305004100016297
[8] Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning. MIT press. · Zbl 1373.68009
[9] Hannig, J., Iyer, H., Lai, R. C., & Lee, T. C. (2016). Generalized fiducial inference: A review and new results. Journal of the American Statistical Association, 111(515), 1346-1361.
[10] Hinton, G. E., & Zemel, R. S. (1994) Autoencoders, minimum description length and helmholtz free energy. In Advances in neural information processing systems, pp. 3-10.
[11] Hornik, K., Stinchcombe, M., & White, H. (1989). Multilayer feedforward networks are universal approximators. Neural networks, 2(5), 359-366. · Zbl 1383.92015
[12] Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
[13] Kingma, D. P., & Welling, M. (2013). Auto‐encoding variational bayes. arXiv preprint arXiv:1312.6114.
[14] Nair, V., & Hinton, G. E. (2010). Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th international conference on machine learning (ICML‐10), pp. 807-814.
[15] Schmidhuber, J (2015). Deep learning i.n neural networks: An overview. Neural networks, 61, 85-117.
[16] Stan Development Team (2020). RStan: the R interface to Stan. R package version 2.19.3.
[17] Tibshirani, R. J., & Efron, B. (1993). An introduction to the bootstrap. Monographs on statistics and applied probability, 57, 1-436. · Zbl 0835.62038
[18] Vincent, P., Larochelle, H., Bengio, Y., & Manzagol, P. A. (2008). Extracting and composing robust features with denoising autoencoders,. In Proceedings of the 25th international conference on Machine learning, ACM, pp. 1096-1103.
[19] Vincent, P., Larochelle, H., Lajoie, I., Bengio, Y., & Manzagol, P. A. (2010). Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. Journal of machine learning research, 11(Dec), 3371-3408. · Zbl 1242.68256
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. In some cases that data have been complemented/enhanced by data from zbMATH Open. This attempts to reflect the references listed in the original paper as accurately as possible without claiming completeness or a perfect matching.