To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time.

4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
Added in 24 Hours
Languages
Recent
Show all languages
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds

Asymptotic theory (statistics)

From Wikipedia, the free encyclopedia

In statistics, asymptotic theory, or large sample theory, is a framework for assessing properties of estimators and statistical tests. Within this framework, it is often assumed that the sample size n may grow indefinitely; the properties of estimators and tests are then evaluated under the limit of n → ∞. In practice, a limit evaluation is considered to be approximately valid for large finite sample sizes too.[1]

YouTube Encyclopedic

  • 1/5
    Views:
    7 182
    6 648
    1 050 078
    2 371
    30 241
  • Asymptotic Normality of OLS parameter Estimators
  • Asymptotic distribution of the maximum likelihood estimator(mle) - finding Fisher information
  • Central limit theorem | Inferential statistics | Probability and Statistics | Khan Academy
  • Jan14-a-AEP-WeakLawOfLargeNumbers
  • Maximum Likelihood - Cramer Rao Lower Bound Intuition

Transcription

Overview

Most statistical problems begin with a dataset of size n. The asymptotic theory proceeds by assuming that it is possible (in principle) to keep collecting additional data, thus that the sample size grows infinitely, i.e. n → ∞. Under the assumption, many results can be obtained that are unavailable for samples of finite size. An example is the weak law of large numbers. The law states that for a sequence of independent and identically distributed (IID) random variables X1, X2, ..., if one value is drawn from each random variable and the average of the first n values is computed as Xn, then the Xn converge in probability to the population mean E[Xi] as n → ∞.[2]

In asymptotic theory, the standard approach is n → ∞. For some statistical models, slightly different approaches of asymptotics may be used. For example, with panel data, it is commonly assumed that one dimension in the data remains fixed, whereas the other dimension grows: T = constant and N → ∞, or vice versa.[2]

Besides the standard approach to asymptotics, other alternative approaches exist:

  • Within the local asymptotic normality framework, it is assumed that the value of the "true parameter" in the model varies slightly with n, such that the n-th model corresponds to θn = θ + h/n . This approach lets us study the regularity of estimators.
  • When statistical tests are studied for their power to distinguish against the alternatives that are close to the null hypothesis, it is done within the so-called "local alternatives" framework: the null hypothesis is H0: θ = θ0 and the alternative is H1: θ = θ0 + h/n . This approach is especially popular for the unit root tests.
  • There are models where the dimension of the parameter space Θn slowly expands with n, reflecting the fact that the more observations there are, the more structural effects can be feasibly incorporated in the model.
  • In kernel density estimation and kernel regression, an additional parameter is assumed—the bandwidth h. In those models, it is typically taken that h → 0 as n → ∞. The rate of convergence must be chosen carefully, though, usually hn−1/5.

In many cases, highly accurate results for finite samples can be obtained via numerical methods (i.e. computers); even in such cases, though, asymptotic analysis can be useful. This point was made by Small (2010, §1.4), as follows.

A primary goal of asymptotic analysis is to obtain a deeper qualitative understanding of quantitative tools. The conclusions of an asymptotic analysis often supplement the conclusions which can be obtained by numerical methods.

Modes of convergence of random variables

Asymptotic properties

Estimators

Consistency

A sequence of estimates is said to be consistent, if it converges in probability to the true value of the parameter being estimated:

That is, roughly speaking with an infinite amount of data the estimator (the formula for generating the estimates) would almost surely give the correct result for the parameter being estimated.[2]

Asymptotic distribution

If it is possible to find sequences of non-random constants {an}, {bn} (possibly depending on the value of θ0), and a non-degenerate distribution G such that

then the sequence of estimators is said to have the asymptotic distribution G.

Most often, the estimators encountered in practice are asymptotically normal, meaning their asymptotic distribution is the normal distribution, with an = θ0, bn = n, and G = N(0, V):

Asymptotic confidence regions

Asymptotic theorems

See also

References

  1. ^ Höpfner, R. (2014), Asymptotic Statistics, Walter de Gruyter. 286 pag. ISBN 3110250241, ISBN 978-3110250244
  2. ^ a b c A. DasGupta (2008), Asymptotic Theory of Statistics and Probability, Springer. ISBN 0387759700, ISBN 978-0387759708

Bibliography

This page was last edited on 23 February 2022, at 13:40
Basis of this page is in Wikipedia. Text is available under the CC BY-SA 3.0 Unported License. Non-text media are available under their specified licenses. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc. WIKI 2 is an independent company and has no affiliation with Wikimedia Foundation.