Preview

Event Will Never Forget

Powerful Essays
Open Document
Open Document
2313 Words
Grammar
Grammar
Plagiarism
Plagiarism
Writing
Writing
Score
Score
Event Will Never Forget
Comparison of Di erent Neighbourhood Sizes in Simulated Annealing
Xin Yao Department of Computer Science University College, University of New South Wales Australian Defence Force Academy Canberra, ACT, Australia 2600

Abstract
Neighbourhood structure and size are important parameters in local search algorithms. This is also true for generalised local search algorithms like simulated annealing. It has been shown that the performance of simulated annealing can be improved by adopting a suitable neighbourhood size. However, previous studies usually assumed that the neighbourhood size was xed during search. This paper presents a simulated annealing algorithm with a dynamic neighbourhood size which depends on the current \temperature" value during search. A method of dynamically deciding the neighbourhood size by approximating a continuous probability distribution is given. Four continuous probability distributions are used in our experiments to generate neighbourhood sizes dynamically, and the results are compared.

combinatorial optimisation. A method of generating dynamic neighbourhood sizes by approximating continuous probability distributions is given in this section. Section 4 compares the experimental results of using di erent continuous probability distributions to generate dynamic neighbourhood sizes. Finally, Section 5 concludes with some remarks and directions of future research.

2 General Simulated Annealing
Although SA can be used in both continuous and discrete cases, this paper only considers combinatorial optimisation by SA unless otherwise indicated explicitly. A combinatorial optimisation problem can be informally described as nding an optimal con guration X from a nite or in nite countable con guration space S . Each con guration X 2 S can be represented by its n (> 0) components, i.e., X = (x1; x2; ; xn ), where xi 2 Xi , i = 1; 2; ; n. An excellent discussion of combinatorial optimisation and its complexity can be found in Garey and



References: 1] P. J. M. van Laarhoven and E. H. L. Aarts, Simulated Annealing: Theory and Applications, D. Reidel Publishing Co., 1987. 2] D. H. Ackley, A Connectionist Machine for Genetic Hillclimbing, Kluwer Academic Publishers, Boston, 1987. 3] X. Yao, Optimization by genetic annealing," In M. Jabri, editor, Proc. of ACNN '91, pages 94{97, Sydney, 1991. 4] D. R. Greening, Parallel simulated annealing techniques," Physica D, 42:293{306, 1990. 5] X. Yao, Simulated annealing with extended neighbourhood," International J. of Computer Math., 40:169{189, 1991. 6] X. Yao and G.-J. Li, General simulated annealing," J. of Computer Sci. & Tech., 6:329{ 338, 1991. 7] S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi, Optimization by simulated annealing," Science, 220:671{680, 1983. 8] M. R. Garey and D. S. Johnson, Computers and Intractability: A Guide to the Theory of NP-Completeness, W.H.Freeman Co., San Francisco, 1979. 9] S. Anily and A. Federgruen, Ergodicity in parameteric nonstationary Markov chains: an application to annealing methods," Oper. Res., 35:867{874, 1987. 10] L. Goldstein and M. Waterman, Neighborhood size in the simulated annealing algorithm," Amer. J. of Math. and Management Sci., 8:409{423, 1988. 11] K. M. Cheh, J. B. Goldberg, and R. G. Askin, A note on the e ect of neighborhood structure in simulated annealing algorithm," Computers and Oper. Res., 18:537{547, 1991. 12] H. H. Szu and R. L. Hartley, Nonconvex optimization by fast simulated annealing," Proc. of IEEE, 75:1538{1540, 1987. 13] W. Feller, An Introduction to Probability Theory and Its Applications, volume 2, John Wiley & Sons, Inc., 2nd edition, 1971. 4 Table 2: SA with a dynamic neighbourhood size which is generated by the Cauchy function (CauSA), Normal function (NorSA), Exponential function (ExpSA), and Stable function with index 1=2 (StableSA). research issue in search theory, i.e., the issue of exploration versus exploitation or global search versus local search. Although local search based on some heuristics can be quite e cient under many circumstances, the problem of local optima is very hard to deal with. Some kind of global search has to be used if a global optimum or near optimum is required. However, the computational cost of global search is often prohibitively high for most real-world applications due to the vast search space. It is bene cial to combine global and local search together. An open question here is how to decide when global or local search should be performed. It is also di cult to draw the line strictly between local and global search in practice. Dynamic neighbourhood size offers a way to deal with the problem by transferring from global search to local search smoothly based on a control parameter, temperature in SA. However, more work has to be done on deciding which kind of generation functions is most suitable for an application, i.e., what is the optimal rate of reducing the neighbourhood size. As indicated before, Fast SA 12] o ers a big improvement over classical SA 7] due to the adoption of Cauchy distribution. An interesting topic is to investigate whether the discrete version of Fast SA can o er similar improvement over classical SA. Our preliminary experiments seem to give a negative answer. Acknowledgement | The author is grateful to Drs. B. Marksjo and R. Sharpe for their support of his work while he was with CSIRO Division of Building, Construction and Engineering.

You May Also Find These Documents Helpful

  • Good Essays

    This figure assumes that the main reference set, covers the indicated circle of sections A, B, C. The solution 1 is created from a convex combination of reference solutions A, B that is added to the reference set as the only solution. In a similar way, combining of convex and non-convex reference of new and original solutions are created points 2, 3 and 4. The complete reference set are including 7 solutions (members) that is shown in the figure above. In genetic algorithm, two solutions are selected randomly from the population and a crossover operator used for the production of one or more children. GA are including a sample population of 100 elements that are selected randomly to create crossover. But in scatter search, two or more of the reference set in a systematic approach in order to produce new…

    • 623 Words
    • 3 Pages
    Good Essays
  • Good Essays

    References: © The Authors JCSCR. (2012). A Comparative Study on the Performance. LACSC – Lebanese Association for Computational Sciences Registered under No. 957, 2011, Beirut, Lebanon, 1-12.…

    • 664 Words
    • 4 Pages
    Good Essays
  • Good Essays

    In this way, the assignment algorithm will converge faster without affecting the network performance. The edgetrimming strategy must follow these design objectives:…

    • 8009 Words
    • 33 Pages
    Good Essays
  • Satisfactory Essays

    LYT2 Task2

    • 4061 Words
    • 12 Pages

    Stein, S. S., Gerding, E. H., Rogers, A. C., Larson, K. K., & Jennings, N. R. (2011). Algorithms…

    • 4061 Words
    • 12 Pages
    Satisfactory Essays
  • Better Essays

    Hi-Ho Yo-Yo, Inc

    • 1077 Words
    • 5 Pages

    References: Hochbaum, D. S. (1999). The Scheduling Problems. Retrieved On October 20, 2011 from riot.ieor.berkeley.edu/riot/Applications/Scheduling/algorithms.html…

    • 1077 Words
    • 5 Pages
    Better Essays
  • Powerful Essays

    This report adheres to the Academic Honesty Statement the professor has on file for all group members.…

    • 2053 Words
    • 11 Pages
    Powerful Essays
  • Powerful Essays

    The aforementioned topics have been modelled to manipulate practical variables in order to achieve the recommended optimal design.…

    • 2213 Words
    • 9 Pages
    Powerful Essays
  • Good Essays

    Many questions arise when we think about optimization problem in healthcare industry. Like how to decide the best location for OPD and emergency vehicles…

    • 523 Words
    • 3 Pages
    Good Essays
  • Satisfactory Essays

    The search starts with creating a random population of grey wolves (candidate solutions) in the GWO algorithm. During the iterations, α, β, and δ estimate the probable position of the prey. Then Each candidate solution updates its position from the prey accordingly. The parameter a is decreased from 2 to 0 in order to emphasize exploration and exploitation, respectively. Candidate solutions diverge from the prey if |A| > 1 and converge towards the prey if |A| < 1. Finally, the GWO algorithm is terminated by the satisfaction of an end…

    • 575 Words
    • 3 Pages
    Satisfactory Essays
  • Powerful Essays

    particular application, the proposed heuristic based on majority rules combined with concessions to significant mi-…

    • 8558 Words
    • 68 Pages
    Powerful Essays
  • Powerful Essays

    Recent developments in optimization techniques that deals in finding the solution of combinatorial optimization problems has provided engineering designers new capabilities. These new optimization algorithms are called metaheuristic techniques and they use nature as a source of inspiration to develop new numerical optimization procedures. It is shown in the literature that these techniques are robust and efficient and their performance is not affected by the complexity of optimization problems. In last two decades several metaheuristic algorithms are developed that mimic natural phenomena. Among these evolutionary algorithms imitate evolutionary biology and make use of the principle of the survival of the fittest to establish a numerical search algorithm. Swarm intelligence is based on the collective behaviour of insect swarm, bird flocking or fish schooling. Particle swarm optimizer turns this collective behaviour of particles into a numerical optimization algorithm. Differential evolution is a method that optimizes a problem by iteratively trying to improve a candidate solution with regard to a given measure of quality. Harmony search method mimics the musical performance process that takes place when a musician searches for a better state of harmony. Big Bang-Big Crunch method simulates the theory of evolution of the universe. Artificial bee colony algorithm is based on the intelligent behaviour of honey bee swarm. Fireflies communicate, search for pray and find mates using bioluminescence with varied flashing patterns. Firefly algorithm mimics the social behaviour of fireflies. Cuckoo search algorithm…

    • 15059 Words
    • 61 Pages
    Powerful Essays
  • Powerful Essays

    Systems-on-a-chip (SOC) are comprised of a collection of pre-designed and preverified cores and user defined logic (UDL). As the complexity of systems-on-a-chip continues to increase, the difficulty and cost of testing such chips is increasing rapidly.…

    • 13186 Words
    • 53 Pages
    Powerful Essays
  • Good Essays

    General method • Useful technique for optimizing search under some constraints • Express the desired solution as an n-tuple (x1 , . . . , xn ) where each xi ∈ Si , Si being a finite set • The solution is based on finding one or more vectors that maximize, minimize, or satisfy a criterion function P (x1 , . . . , xn ) • Sorting an array a[n] – Find an n-tuple where the element xi is the index of ith smallest element in a – Criterion function is given by a[xi ] ≤ a[xi+1 ] for 1 ≤ i < n – Set Si is a finite set of integers in the range [1,n] • Brute force approach – Let the size of set Si be mi – There are m = m1 m2 · · · mn n-tuples that satisfy the criterion function P – In brute force algorithm, you have to form all the m n-tuples to determine the optimal solutions • Backtrack approach – Requires less than m trials to determine the solution – Form a solution (partial vector) and check at every step if this has any chance of success – If the solution at any point seems not-promising, ignore it – If the partial vector (x1 , x2 , . . . , xi ) does not yield an optimal solution, ignore mi+1 · · · mn possible test vectors even without looking at them • All the solutions require a set of constraints divided into two categories: explicit and implicit constraints Definition 1 Explicit constraints are rules that restrict each xi to take on values only from a given set. – Explicit constraints depend on the particular instance I of problem being solved – All tuples that satisfy the explicit constraints define a possible solution space for I – Examples of explicit constraints ∗ xi ≥ 0, or all nonnegative real numbers ∗ xi = {0, 1} ∗ li ≤ xi ≤ ui Definition 2 Implicit constraints are rules that determine which of the tuples in the solution space of I satisfy the criterion function. – Implicit constraints describe the way in which the xi s must relate to each other. • Determine problem solution by systematically searching the solution space for the given problem instance…

    • 1196 Words
    • 5 Pages
    Good Essays
  • Powerful Essays

    This chapter provides a breif introduction to the area of Nature Inspired Algorithms and its classification. Among all the categories, Swarm intelligence will be the main area of focus. Various swarm intelligence techinques will be discussed in the later sections. Cuckoo Search Algorithm is also one of the latest swarm intelligence technique which is focused in this dissertation to solve the problem discussed in chapter 1.…

    • 3691 Words
    • 15 Pages
    Powerful Essays
  • Good Essays

    Many scientists have been provided more different kinds of search plan fit the nature of the search area. Sometimes we search for the targets in difficult terrain areas on the surface of the ground or in the deep of the sea. Specialists in this field interested in dividing those areas to a set of identical or different states in order to increase the probability of detection or minimize the search effort. The nature of the search area is control in the form of the cells. Hong et al. [1-2] divided the area into hexagonal cells. They proposed an approximation algorithm for the optimal search path. This algorithm optimizes an approximate path to compute detection probability by using the conditional probabilities. Then, finding the maximum detection probability search path. Song and Teneketizs [3] determined the optimal search strategies with multiple sensors that maximize the total probability of successful search where the target is hidden in one of a finite set of different cells. Teamah et al. [4] divided the search region into square cells. They minimized the probability of undetection and the searching effort (is bounded by a normal distribution) by using multiple searchers. They studied some special cases when the target is hidden in one of M-identical cells and when the effort is…

    • 1295 Words
    • 6 Pages
    Good Essays