NUST Institutional Repository

Large Sample Techniques for Statistics

Show simple item record

dc.contributor.author Jiming Jiang
dc.date.accessioned 2023-12-06T04:52:58Z
dc.date.available 2023-12-06T04:52:58Z
dc.date.issued 2010
dc.identifier.isbn 978-1-4419-6826-5
dc.identifier.uri http://10.250.8.41:8080/xmlui/handle/123456789/40919
dc.description.abstract In a way, the world is made up of approximations, and surely there is no exception in the world of statistics. In fact, approximations, especially large sample approximations, are very important parts of both theoretical and applied statistics. The Gaussian distribution, also known as the normal distribution, is merely one such example, due to the well-known central limit theorem. Large-sample techniques provide solutions to many practical problems; they simplify our solutions to difficult, sometimes intractable problems; they justify our solutions; and they guide us to directions of improvements. On the other hand, just because large-sample approximations are used everywhere, and every day, it does not guarantee that they are used properly, and, when the techniques are misused, there may be serious consequences. Example 1 (Asymptotic χ2 distribution). Likelihood ratio test (LRT) is one of the fundamental techniques in statistics. It is well known that, in the “standard” situation, the asymptotic null distribution of the LRT is χ2, with the degrees of freedom equal to the difference between the dimensions, defined as the numbers of free parameters, of the two nested models being compared (e.g., Rice 1995, pp. 310). This might lead to a wrong impression that the asymptotic (null) distribution of the LRT is always χ2. A similar mistake might take place when dealing with Pearson’s χ2-test—the asymptotic distribution of Pearson’s χ2-test is not always χ2 (e.g., Moore 1978). Example 2 (Approximation to a mean). It might be thought that, in a large sample, one could always approximate the mean of a random quantity by the quantity itself. In some cases this technique works. For example, suppose X1,...,Xn are observations that are independent and identically distributed (i.i.d.) such that μ = E(X1) = 0. Then one can approximate E(n i=1 Xi) = nμ by simply removing the expectation sign, that is, by n i=1 Xi. This is because the difference n i=1 Xi − nμ = n i=1(Xi − μ) is of the order O( √n), which is lower than the order of the mean of n i=1 Xi. However, this technique completely fails if one considers (n i=1 Xi)2 instead. To see this, let us assume for simplicity that Xi ∼ N(0, 1). Then E(n i=1 Xi)2 = n. On the other hand, since n i=1 Xi ∼ N(0, n), (n i=1 Xi)2 = n{(1/ √n) n i=1 Xi}2 ∼ nχ2 1, where en_US
dc.language.iso en en_US
dc.publisher Springer en_US
dc.title Large Sample Techniques for Statistics en_US
dc.type Book en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account