To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time. 4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds VEGAS algorithm

The VEGAS algorithm, due to G. Peter Lepage, is a method for reducing error in Monte Carlo simulations by using a known or approximate probability distribution function to concentrate the search in those areas of the integrand that make the greatest contribution to the final integral.

The VEGAS algorithm is based on importance sampling. It samples points from the probability distribution described by the function $|f|,$ so that the points are concentrated in the regions that make the largest contribution to the integral.

• 1/3
Views:
2 455
34 843
20 676
• Las Vegas Algorithm | Randomized Algorithm
• Randomized algorithms (intro) | Journey into cryptography | Computer Science | Khan Academy
• Randomized Quick Sort - Algorithms Design & Analysis

Sampling method

In general, if the Monte Carlo integral of $f$ is sampled with points distributed according to a probability distribution described by the function $g,$ we obtain an estimate $\mathrm {E} _{g}(f;N),$ $\mathrm {E} _{g}(f;N)={1 \over N}\sum _{i}^{N}{f(x_{i})}/g(x_{i}).$ The variance of the new estimate is then

$\mathrm {Var} _{g}(f;N)=\mathrm {Var} (f/g;N)$ where $\mathrm {Var} (f;N)$ is the variance of the original estimate, $\mathrm {Var} (f;N)=\mathrm {E} (f^{2};N)-(\mathrm {E} (f;N))^{2}.$ If the probability distribution is chosen as $g=|f|/I(|f|)$ then it can be shown that the variance $\mathrm {Var} _{g}(f;N)$ vanishes, and the error in the estimate will be zero. In practice it is not possible to sample from the exact distribution g for an arbitrary function, so importance sampling algorithms aim to produce efficient approximations to the desired distribution.

Approximation of probability distribution

The VEGAS algorithm approximates the exact distribution by making a number of passes over the integration region while histogramming the function f. Each histogram is used to define a sampling distribution for the next pass. Asymptotically this procedure converges to the desired distribution. In order to avoid the number of histogram bins growing like $K^{d}$ with dimension d the probability distribution is approximated by a separable function: $g(x_{1},x_{2},\ldots )=g_{1}(x_{1})g_{2}(x_{2})\cdots$ so that the number of bins required is only Kd. This is equivalent to locating the peaks of the function from the projections of the integrand onto the coordinate axes. The efficiency of VEGAS depends on the validity of this assumption. It is most efficient when the peaks of the integrand are well-localized. If an integrand can be rewritten in a form which is approximately separable this will increase the efficiency of integration with VEGAS.