To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time.

4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
Added in 24 Hours
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds

BRST algorithm

From Wikipedia, the free encyclopedia

Boender-Rinnooy-Stougie-Timmer algorithm (BRST) is an optimization algorithm suitable for finding global optimum of black box functions. In their paper Boender et al. [1] describe their method as a stochastic method involving a combination of sampling, clustering and local search, terminating with a range of confidence intervals on the value of the global minimum.

The algorithm of Boender et al. has been modified by Timmer.[2] Timmer considered several clustering methods. Based on experiments a method named "multi level single linkage" was deemed most accurate.

Csendes' algorithms [3] are implementations of the algorithm of [Boender et al.][1] and originated the public domain software product GLOBAL. The local algorithms used are a random direction, linear search algorithm also used by Törn, and a quasi—Newton algorithm not using the derivative of the function. The results show the dependence of the result on the auxiliary local algorithm used.

Background

Extending the class of functions to include multimodal functions makes the global optimization problem unsolvable in general. In order to be solvable some smoothness condition on the function in addition to continuity must be known.

The existence of several local minima and unsolvability in general are important characteristics of global optimization. Unsolvability here means that a solution cannot be guaranteed in a finite number of steps. There are two ways to deal with the unsolvability problem. First, "a priori" conditions on f and A are posed which turns the problem into a solvable one or at least makes it possible to tell for sure that a solution has been found. This restricts the function class that is considered. The second approach which allows a larger class of objective functions to be considered is to give up the solvability requirement and only try to obtain an estimate of the global minimum. In this "probabilistic" approach it would be desirable also to obtain some results on the goodness of an obtained estimate. Some of the solvable problems may fall in this class because the number of steps required for a guaranteed solution might be prohibitively large.

When relaxing the requirement on solvability it seems rational to require that the probability that a solution is obtained approaches 1 if the procedure is allowed to continue forever. An obvious probabilistic global search procedure is to use a local algorithm starting from several points distributed over the whole optimization region. This procedure is named "Multistart". Multistart is certainly one of the earliest global procedures used. It has even been used in local optimization for increasing the confidence in the obtained solution. One drawback of Multistart is that when many starting points are used the same minimum will eventually be determined several times. In order to improve the efficiency of Multistart this should be avoided.

Clustering methods are used to avoid this repeated determination of local minima. This is realized in three steps which may be iteratively used. The three steps are:

  • (a) Sample points in the region of interest.
  • (b) Transform the sample to obtain points grouped around the local minima.
  • (c) Use a clustering technique to recognize these groups (i.e. neighbourhoods of the local minima).

If the procedure employing these steps is successful then starting a single local optimization from each cluster would determine the local minima and thus also the global minimum. The advantage in using this approach is that the work spared by computing each minimum just once can be spent on computations in (a) and (b), which will increase the probability that the global minimum will be found.

Being a clustering method, their effectiveness is higher for low-dimensional problems and become less effective for problems having a few hundred variables.

References

  1. ^ a b Boender, C.G.E.; A.H.G. Rinnooy Kan; L. Strougie; G.T. Timmer (1982). "A stochastic method for global optimization" (PDF). Mathematical Programming. 22: 125–140. doi:10.1007/BF01581033. S2CID 5450000.
  2. ^ Timmer, G.T. (1984). Global optimization: A stochastic approach (Ph.D. Thesis). Erasmus University Rotterdam.
  3. ^ Csendes, T. (1988). "Nonlinear parameter estimation by global optimization—Efficiency and reliability". Acta Cybernetica. 8 (4): 361–370.

External links

  • http://www.abo.fi/~atorn/Globopt.html With the author's permission, text has been verbatim copied.
  • Janka Compares various global optimization algorithms, of which BRST shows superior performance.
  • Janka Presents the number of function-evaluations performed on the testset of Dixon-Szegö. Along with the MCS algorithm, the BRST requires the lowest number of evaluations.
This page was last edited on 18 February 2024, at 07:49
Basis of this page is in Wikipedia. Text is available under the CC BY-SA 3.0 Unported License. Non-text media are available under their specified licenses. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc. WIKI 2 is an independent company and has no affiliation with Wikimedia Foundation.