The focus of this book is on the birth and historical development of permutation statistical methods from the early 1920s to the near present. Beginning with the seminal contributions of R.A. Fisher, E.J.G. Pitman, and others in the 1920s and 1930s, permutation statistical methods were initially introduced to validate the assumptions of classical statistical methods. Permutation methods have advantages over classical methods in that they are optimal for small data sets and non-random samples, are data-dependent, and are free of distributional assumptions. Permutation probability values may be exact, or estimated via moment- or resampling-approximation procedures. Because permutation methods are inherently computationally-intensive, the evolution of computers and computing technology that made modern permutation methods possible accompanies the historical narrative. Permutation analogs of many well-known statistical tests are presented in a historical context, including multiple correlation and regression, analysis of variance, contingency table analysis, and measures of association and agreement. A non-mathematical approach makes the text accessible to readers of all levels.Math. Comput. 15, 192a195 (1961) Wells, M.B.: Computing at LASL in the 1940s and 1950s: MANIAC. Tech. rep., Los Alamos Scientific Laboratory, Los ... Accessed 17 Jan 2012 Wilson, E.B.: The controlled experiment and the four-fold table.
|Title||:||A Chronicle of Permutation Statistical Methods|
|Author||:||Kenneth J. Berry, Janis E Johnston, Paul W. Mielke|
|Publisher||:||Springer Science & Business Media - 2014-04-11|