亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        Rock Hyraxes Swarm Optimization:A New Nature-Inspired Metaheuristic Optimization Algorithm

        2021-12-14 09:58:54BelalAlKhateebKawtherAhmedMahaMahmoodandDacNhuongLe
        Computers Materials&Continua 2021年7期

        Belal Al-Khateeb,Kawther Ahmed,Maha Mahmood and Dac-Nhuong Le

        1College of Computer Science and Information Technology,University of Anbar,Ramadi,Iraq

        2General Directorate of Scientific Welfare,Ministry of Youth and Sport,Baghdad,Iraq

        3Institute of Research and Development,Duy Tan University,Danang,550000,Vietnam

        4Faculty of Information Technology,Duy Tan University,Danang,550000,Vietnam

        Abstract: This paper presents a novel metaheuristic algorithm called Rock Hyraxes Swarm Optimization (RHSO) inspired by the behavior of rock hyraxes swarms in nature.The RHSO algorithm mimics the collective behavior of Rock Hyraxes to find their eating and their special way of looking at this food.Rock hyraxes live in colonies or groups where a dominant male watch over the colony carefully to ensure their safety leads the group.Forty-eight(22 unimodal and 26 multimodal)test functions commonly used in the optimization area are used as a testing benchmark for the RHSO algorithm.A comparative efficiency analysis also checks RHSO with Particle Swarm Optimization (PSO), Artificial-Bee-Colony (ABC), Gravitational Search Algorithm(GSA),and Grey Wolf Optimization(GWO).The obtained results showed the superiority of the RHSO algorithm over the selected algorithms; also, the obtained results demonstrated the ability of the RHSO in convergence towards the global optimal through optimization as it performs well in both exploitation and exploration tests.Further, RHSO is very effective in solving real issues with constraints and new search space.It is worth mentioning that the RHSO algorithm has a few variables,and it can achieve better performance than the selected algorithms in many test functions.

        Keywords: Optimization; metaheuristic; constrained optimization; rock hyraxes swarm optimization; RHSO

        1 Introduction

        Nature is full of social behaviors that work to accomplish many different tasks.Naturally, all these behaviors coexist in different environments, and their main goal is to survive, but there is what makes them work in the form of swarms, groups, flocks, and colonies because of hunting and mobility and defend themselves.The food search is also essential for social interactions that allow them to complete their lives and reproduction.Another reason for the swarm for some creature is navigation.Birds are the best examples of such behaviors, which migrate intercontinental in flocks appropriately [1,2].

        Metaheuristic optimization techniques are used extensively to solve most of the problems;this made it very common to be used as primary methods of acquiring the optimum results of optimization problems.Especially, few of them like Genetic Algorithm (GA) [3], Grey Wolf Optimization (GWO) [4], Ant Colony Optimization (ACO) [5], Slap Swarm Optimization (SSO) [6],Artificial-Bee-Colony (ABC) [7] and Particle Swarm Optimization (PSO) [8] are very known and used in broad distinct fields.Metaheuristics have become very common for several reasons:metaheuristics are very clear.They are fundamentally inspired by physical phenomena, animal behaviors, or evolutionary concepts.Metaheuristics are easy to use, which allow researchers to mimic natural behavior, enhance or produce new metaheuristics and hybridize more than one metaheuristics.Furthermore, this simplicity helps scientists to acquire information rapidly and adapt them for problems solving.Also, flexibility indicates the applicability of metaheuristics to various problems without notable changes in the algorithm’s structure.Metaheuristics can be easily applied to various problems.Thus, all a designer need is knowing and understanding how to represent his problems.Also, most metaheuristics have derivation-free mechanisms.In inequality to gradient-based optimization approaches, metaheuristics enhance problems randomly.Metaheuristics deals with problems stochastically.The optimization process starts with random solutions,and there is no need to calculate the derivative of search spaces to find the optimal solution.Finally, metaheuristics are more capable than classical optimization strategies in managing local optima.This is due to the random nature of metaheuristics, which permit them to overcome local solutions inactivity and widely search the whole search area.

        Swarm Intelligence (SI) was proposed in 1993 [9] and has been defined by Bonabeau et al.[3]as “the emergent collective intelligence of groups of simple agents”.The SI inspiration techniques are created primarily from natural habitations, herds, and flock.There are many examples of SI popular techniques, such as PSO [8], ABC [7], and ACO [5].

        Regardless of the variation among metaheuristics, usually, they have a common trait, which is dividing the search process into the exploration and exploitation phases [10–12].The exploration phase indicates the investigating process of search space as widely as possible.An algorithm must have the stochastic operators globally and randomly search within the search space to support this phase.In contrast, exploitation indicates local search capability about the concerned regions taken from the previous phase (exploration).The appropriate balance of these phases is seen as a challenging task because of the metaheuristics stochastically nature.This paper introduces a new SI technique that is inspired by the behavior of Rock Hyraxes swarms.This algorithm has few parameters that make it easy to implement; the proposed algorithm can also balance between exploration and exploitation phases, making it appropriate for solving many optimization problems.The rest of the paper is organized as follows:Section 2 shows the relevant works.Section 3 presents the Rock Hyraxes Swarm Optimization (RHSO) Algorithm inspiration and mathematical model.Section 4 presents the results and discussion of test functions.Finally,Section 5 concludes the work and suggests several future research directions.

        2 Literature Review

        There are three major categories of metaheuristics; those are Evolutionary Algorithm (EA),Physics-Based (PB), and Swarm Intelligence (SI) algorithms.EAs, which inspired by the genetic and evolutionary behaviors of creatures.In this branch, the most popular algorithm is the Genetic Algorithm (GA).GA was proposed by Holland in 1992 [13] to simulate Darwinian evolution concepts.Goldberg [14] extensively investigated the engineering applications of GA.Typically,the development is finished by evolving associate initial random solution in EAs.Each new population is created by the cross over and mutation of the previous generation solutions.Since the best solution can higher generate the new population, which most likely creates this new population higher than the population within the previous generation(s), this may make the initial random population converges to the best solutions over generations.An example of EAs are Differential Evolution (DE) [15], Evolutionary Programming (EP) [16], Evolution Strategy(ES) [17,18], Genetic Programming (GP) [19] and Biogeography Based Optimizer (BBO) [20].Simon initially produced the BBO algorithm rule in 2008 [20].The elemental set up of this algorithm program is galvanized by biology science that refers to biological organisms’study in geographical distribution.

        In physics-based techniques, the development algorithms generally simulate the physical rules.Among the physics-based metaheuristic algorithms are Gravitated Native Search (GLSA) [21], Big Bang Big Crunch (BBBC) [22], Gravitated Search Algorithm (GSA) [23], Charged System Search(CSS) [24], Artificial Reaction Improvement Rule [25], Part rule [26], Ray improvement [27] rule,Small World improvement rule [28], Galaxy based Search Algorithm (GbSA) [29] and arced area improvement [30].Those algorithms’method is completely unlike EAs as it uses a random set of search spaces that can communicate and move throughout search space per physical rules.This movement is enforced, for instance, exploitation gravitation, inertia force, magnetic force, weights,ray casting, and many others.

        GSA is an example of a physics-based algorithm rule at that its basic physical theory is impressed by Newton’s law of universal gravitation.The GSA performs a search by employing a group of agents with lots of proportional to the value of fitness performance.Throughout iteration, the gravitated forces between them interest the masses in each another [23].

        Mirjalili et al.[4] propose a new metaheuristic called Grey Wolf Optimizer (GWO) inspired by grey wolves’behavior.The GWO algorithmic simulates the leadership hierarchy and looking mechanism of grey wolves in nature.Four grey wolves, such as alpha, delta, beta, and omega, are used to simulate the leadership hierarchy.Additionally, GWO implements three significant steps of looking; those are hunting, attacking prey, and peripheral prey.

        SI methods are algorithms that simulate the social behavior of flocks, herds, swarms, or other creatures.The mechanism is near to the performance of the physics-based mostly rule; but the search agents explore looking space and simulate creatures’collective and social intelligence.PSO proposed by Kennedy and Eberhart [8] is one of the famous SI techniques samples.PSO is inspired by the birds flocking social behavior.PSO uses multiple particles that follow the most effective particle’s position and their own to data best-obtained positions.

        The last example of SI is the Bees’swarming.Artificial Bee Colony Algorithm introduced by Dervis Karaboga et al.[31] in (2007), ABC is an associate optimization algorithm supported the intelligent behavior of bee swarm.ABC testing on tests and compare it with different algorithms like PSO and GA.The results showed that the ABC algorithm outperforms the opposite algorithms.

        From all the above, it can be seen that no one metaheuristics algorithm can solve all kinds of problem domains; therefore, there is always a need for new algorithms that can address many types of problem domains.

        3 Rock Hyraxes Swarm Optimization(RHSO)

        The mathematical model of the proposed method, together with its inspiration, is discussed in this section.

        3.1 Inspiration

        Rock hyrax (Procavia capensis) is a small furry mammal that lives in rocky landscapes across sub-Saharan Africa and along the Arabian Peninsula coast.These mammals live in colonies usually dominated by a single male who aggressively defends his territory; the colonies sometimes up to 50 individuals.They sleep together, look for food together, and even raise their babies together (who then all play together).There are three types of hyrax, two are known as the rock(or bush) hyrax and the third as tree hyrax.In the field, it is sometimes difficult to differentiate between them [32,33].

        Rock hyraxes live in areas that vary widely in ambient temperatures and provide adequate water and feed.Low metabolic rates and transparent body temperatures may have contributed to the successful extraction of the rocky protrusions isolated through their distribution [34].

        The rock hyrax feeds every day in a circle formation, with their head pointing to the outside of the circle to keep an eye out for predators, such as leopards, hyenas, jackals, servals, pythons,and the Verreaux’s eagle and black eagle.When the group is feeding or basking, either the breeding male or a female (Leader) will keep a lookout from a high rock or branch and will give a sharp alarm or call if danger threatens, at which point the group will scurry for cover [35].

        3.2 Mathematical Model and Algorithm

        Rock hyraxes start taking solar baths for several hours and sharing places to live together;they looking for food together in a distinctive way:forming a circle with different dimensions and angles to get their food.When they find food, the Leader takes a higher place to find food and protect each other from predatory animals.

        To mathematically model the Rock hyrax swarm, the population first consists of Leader and members.As mentioned above, the Leader chooses the higher and best place to observe the rest of the group.

        The Leader then updates his location based on his old location using the following equation:

        wherer1refers to a random number between [0,1],xrefers to the previous position of Leader,Leaderposrepresent “old position of the leader” andjrefers to “refers to each diminution.” After updating the leader position, all members (or search agents) update their position based on their position.

        wherecircrefers to circular motion and try to mimic the circle system, it is calculated as:

        n1=r2?cos(ang)

        n2=r2?sin(ang)

        wherer2refers to the radius, and it is a random number between [0, 1],angis a random number between [0, 360], and it refers to the angle of a move.Theangalso updated in every generation,and this update depending on the lower and upper bounds of the variables, where lb and ub are the lower and upper bands of the random number generation.

        The angle (ang) can be kept within the specified range by making it either equal to 360 if the value of the output becomes greater than 360, or equal to 0 if it becomes less than 0.

        The pseudo-code of the RHSO algorithm is shown in Algorithm 1.The RHSO starts optimization by creating random solutions in the explorative mode and calculating their fitness.Depending on their fitness, it identifies the best fitness as Leader; this represents switching from explorative mode to local exploitation mode that focuses on the promising regions when global optimal may be in a close place.Later, the Leader represents the best solution for optimization problems.The search agents start again with another set of explorative moves and subsequent turn into a new exploitation stage.The Leader updates his position based on Eq.(1), while the remaining members update their positions according to the Leader’s position, as shown in Eq.(2).Calculating the fitness of each search agent and selecting the best one as a leader.This process will continue throughout iterations; when reaching the stopping condition, the process returned the Leader as the best approximation for the optimization problem’s best solution.

        ?

        4 Results and Discussion

        RHSO algorithm is evaluated by using 48 benchmark functions.Some of those functions are standard functions that are used in researches.These functions are chosen to be able to show the performance of RHSO and also to compare it with some known algorithms.The selected 48 test functions are shown in Tabs.1 and 2, where D means the function’s dimension, range is the function’s search space limits, and Opt is the optimal value.The selected functions are unimodal or multimodal benchmark minimization functions.Unimodal test functions have a single optimum value; thus, they can benchmark an algorithm’s convergence and exploitation.While multimodal test functions have more than one optimum value, making them more challenging than unimodal.An algorithm should avoid all the local optima to approach and approximate the global optimum.So, exploration and local optima avoidance of algorithms can be benchmarked by multimodal test functions.

        Table 1:Unimodal benchmark functions

        Table 2:Multimodal benchmark functions

        For each benchmark function, the RHSO algorithm and the compared algorithms are performed in the experiments under the condition of the same number of iterations (1000), independent runs for 30 times, and the population size is set to 50.The statistical results (average and standard deviation) are showed in Tabs.3 and 4.For verifying the results, the RHSO algorithm is compared with ABC [7], PSO [8], GSA [23], and GWO [4].

        Table 3:Results of unimodal benchmark functions

        Table 4:Results of multimodal benchmark functions

        4.1 Exploitation Analysis

        The results in Tab.3 demonstrated that RHSO is better than the selected algorithms in all unimodal test functions.Unimodal functions test the exploitation of an algorithm.The obtained results showed RHSO superiority in exploiting the optimal value, so RHSO provides excellent exploitation ability.

        4.2 Exploration Analysis

        For testing the exploration strength of an algorithm, the multimodal functions are used as the number growing exponentially with dimension such types of functions.The results in Tab.4 demonstrated that RHSO is better than the selected algorithms on most multimodal functions.The obtained results show the superiority of the RHSO algorithm in terms of exploration.

        The multimodal results presented in Tab.4 demonstrated the exploitation of the proposed algorithm.Those results show the efficiency and strength of RHSO as compared to the other algorithms that are adopted for comparison.RHSO is tested on 26 multimodal test functions,at which RHSO is better and more efficient than the other algorithms.Also, RHSO can control exploitation better than other algorithms.

        Based on the results in Tab.3, the RHSO algorithm obtained the optimal value in all test functions; Therefore, RHSO has surpassed the selected algorithms, and this makes the RHSO algorithm better than they do in exploitation convergence.Those results demonstrated the accuracy, efficiency, and flexibility of the proposed algorithm.

        According to the results in Tab.4, the RHSO algorithm obtained the optimal value in 23 test functions; also, it is very close to the optimal value and better than the values of the selected algorithms for the same test functions in three test functions (F23, F24, and F28).

        5 Conclusions and Future Work

        This paper proposed a new optimization algorithm inspired by Rock hyrax’s behavior; RHSO is proposed as an alternative technique for solving optimization problems.In the proposed RHSO algorithm, the updating of position makes the solutions move towards or outwards the goal to guarantee the search space exploitation and exploration.Forty-eight test functions are used to test the performance of RHSO in terms of exploitation and exploration.The obtained results demonstrated that RHSO was able to outperform ABC, PSO, GSA, and GWO.The obtained unimodal test functions result demonstrated RHSO algorithm exploitation superiority.After that,RHSO exploration ability is shown by the obtained multimodal test functions result.RHSO algorithm is characterized by having few variables than the other algorithms, making it easy to understand and implement.RHSO can solve various optimization problems (like scheduling,maintenance, parameters optimization, and many others).Future studies can recommend several directions, such as solving various optimization problems and developing a multi-objective version of the RHSO algorithm as RHSO is currently a single objective optimization algorithm.

        Funding Statement:The author(s) received no specific funding for this study.

        Conficts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

        色综合999| 欧洲女人性开放免费网站| 色爱区综合五月激情| 国产精品高潮av有码久久| 国产精品午夜福利亚洲综合网| 好大好爽我要高潮在线观看| 亚洲精品熟女国产| 亚洲手机国产精品| 免费人成黄页网站在线观看国产 | 亚洲av产在线精品亚洲第三站| 免费不卡在线观看av| 免费黄色电影在线观看| 久久99久久99精品免视看国产成人| 亚洲熟女av在线观看| 国产精品国产三级国产专播| 国产精品免费久久久久影院 | 91久久精品一二三区蜜桃| 日本a级黄片免费观看| 国产激情久久久久久熟女老人av | 日夜啪啪一区二区三区| 欧美在线不卡视频| 亚洲国产精品成人av| 国产一区二区自拍刺激在线观看| 无遮无挡爽爽免费毛片| 欧美亚洲国产人妖系列视| 亚洲天堂av在线免费播放| 国产精品理论片在线观看| 国产中老年妇女精品| 久久人妻av无码中文专区| 国产无套一区二区三区久久| 少妇高潮流白浆在线观看| 国产女人18一级毛片视频| 麻豆成人久久精品二区三区91| 麻豆国产精品久久人妻| 亚洲国产成人久久一区www| 亚洲成精品动漫久久精久| 人妻有码av中文幕久久| 天天躁日日躁狠狠躁| 国产亚洲欧美在线观看的| 国产av黄色一区二区| 美女张开腿黄网站免费|