Tytuł artykułu
Autorzy
Identyfikatory
Warianty tytułu
Języki publikacji
Abstrakty
Unbiased estimation appeared to be an accepted golden standard of statistical analysis ever until the Stein's discovery of a surprising phenomenon attributable to multivariate spaces. So called Stein's paradox arises in estimating the mean of a multivariate standard normal random variable. Stein showed that both natural and intuitive estimate of a multivariate mean given by the observed vector itself is not even admissible and may be improved upon under the squared-error loss when the dimension is greater or equal to three. Later Stein and his student James developed so called 'James–Stein estimator', a shrunken estimate of the mean, which had uniformly smaller risk for all values in the parameter space. The paradox first appeared both unintuitive and even unacceptable, but later it was recognised as one of the most influential discoveries of all times in statistical science. Today the 'shrinkage principle' literally permeates the statistical technology for analysing multivariate data, and in its application is not exclusively confined to estimating the mean, but also the covariance structure of multivariate data. We develop shrinkage versions of both the linear and quadratic discriminant analysis and apply them to sparse multivariate gene expression data obtained at the Centre for Biomedical Informatics (CBI) in Prague.
Wydawca
Czasopismo
Rocznik
Tom
Strony
64--67
Opis fizyczny
Bibliogr. 10 poz., tab.
Twórcy
autor
- Institute of Computer Science, Academy of Sciences of the Czech Republic, Pod Vodárenskou věží 2, 182 07 Prague, Czech Republic
autor
- Institute of Computer Science, Academy of Sciences of the Czech Republic, Pod Vodárenskou věží 2, 182 07 Prague, Czech Republic
Bibliografia
- [1] Stein C. Inadmissibility of the usual estimator for the mean of a multivariate distribution. In: Neyman J, editor. Proc. 3rd. Berkeley Symp. Math. Statist. Probab. vol. 1. Univ. California Press; 1956. pp. 197–206.
- [2] James W, Stein C. Estimation with quadratic loss. Proc. 4th Berkeley Symp. Math. Statist. Prob. 1961. vol. 1, pp. 361–79.
- [3] Baranchik A. Multiple regression and estimation of the mean of a multivariate normal distribution, Tech. Rep. 51. Stanford, CA, U.S.A.: Department of Statistics, Stanford University; 1964.
- [4] Bock M. Minimax estimators of the mean of a multivariate normal distribution. Ann Stat 1975;3:209–18.
- [5] Schäfer J, Strimmer K. A shrinkage approach to large-scale covariance matrix estimation and implications for functional genomics. Stat Appl Genet Mol Biol 2005;4. http://dx.doi.org/10.2202/1544-6115.1175.
- [6] Ledoit O, Wolf M. Improved estimation of the covariance matrix of stock returns with an application to portfolio selection. J Empir Finance 2003;10:603–21. http://dx.doi.org/10.1016/s0927-5398(03)00007-0.
- [7] Mahalanobis PC. On the generalised distance in statistics. Proc Natl Inst Sci India 1936;2:49–55, http://www.new.dli.ernet.in/rawdataupload/upload/insa/ INSA_1/20006193_49.pdf.
- [8] Tibshirani R, Hastie T, Narasimhan B, Chu G. Diagnosis of multiple cancer types by shrunken centroids of gene expression. Proc Natl Acad Sci U S A 2002;99(10):6567–72. http://dx.doi.org/10.1073/pnas.082099299.
- [9] Smyth GK. Linear models and empirical Bayes methods for assessing differential expression in microarray experiments. Stat Appl Genet Mol Biol 2004;3(1). http://dx.doi.org/10.2202/1544-6115.1027.
- [10] R Development Core Team. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing; 2008, ISBN: 3-900051-07-0, http://www.R-project.org.
Typ dokumentu
Bibliografia
Identyfikator YADDA
bwmeta1.element.baztech-4966d0b0-0405-426e-bec3-e16e3adafa2a