دانلود رایگان مقاله انگلیسی بهینه ساز گرگ خاکستری به همراه ترجمه فارسی
عنوان فارسی مقاله | بهینه ساز گرگ خاکستری |
عنوان انگلیسی مقاله | Grey Wolf Optimizer |
رشته های مرتبط | مهندسی کامپیوتر، مهندسی نرم افزار، مهندسی الگوریتم ها و محاسبات |
کلمات کلیدی | بهینه سازی، تکنیک های بهینه سازی، الگوریتم اکتشافی، متهوریستی، بهینه سازی محدود، GWO |
فرمت مقالات رایگان |
مقالات انگلیسی و ترجمه های فارسی رایگان با فرمت PDF آماده دانلود رایگان میباشند همچنین ترجمه مقاله با فرمت ورد نیز قابل خریداری و دانلود میباشد |
کیفیت ترجمه | کیفیت ترجمه این مقاله متوسط میباشد |
نشریه | الزویر – Elsevier |
مجله | پیشرفت در نرم افزار مهندسی – Advances in Engineering Software |
سال انتشار | 2014 |
کد محصول | F547 |
مقاله انگلیسی رایگان (PDF) |
دانلود رایگان مقاله انگلیسی |
ترجمه فارسی رایگان (PDF) |
دانلود رایگان ترجمه مقاله |
خرید ترجمه با فرمت ورد |
خرید ترجمه مقاله با فرمت ورد |
جستجوی ترجمه مقالات | جستجوی ترجمه مقالات مهندسی کامپیوتر |
فهرست مقاله: چکیده 1. مقدمه 2- مرور منابع 3- بهینه ساز گرگ خاکستری(GWO) 3-1 الهام 3-2-1 سلسله مراتب اجتماعی 3-2-2 محاصره شکار 3-2-3 شکار 3-2-4 حمله به شکار( بهره برداری) 3-2-5 جست وجوی شکار(اکتشاف) 4- نتایج و بحث 4-1 تحلیل بهره برداری 4-2 تحلیل اکتشاف 4-3 اجتناب از حداقل محلی 4-4 تحلیل رفتار همگرایی(نزدیک شدن) 5-GWO برای مسائل مهندسی کلاسیک 5-1 طراحی فنر کشش/ فشار 5-2 طراحی تیر جوش یافته 5-3 طرح محفظه تحت فشار 6- کاربرد واقعی GWO در مهندسی اپتیک و لیزر( طراحی بافر اپتیکی) 7- نتیجه گیری |
بخشی از ترجمه فارسی مقاله: 1. مقدمه |
بخشی از مقاله انگلیسی: 1. Introduction Meta-heuristic optimization techniques have become very popular over the last two decades. Surprisingly, some of them such as Genetic Algorithm (GA) [1], Ant Colony Optimization (ACO) [2], and Particle Swarm Optimization (PSO) [3] are fairly well-known among not only computer scientists but also scientists from different fields. In addition to the huge number of theoretical works, such optimization techniques have been applied in various fields of study. There is a question here as to why meta-heuristics have become remarkably common. The answer to this question can be summarized into four main reasons: simplicity, flexibility, derivation-free mechanism, and local optima avoidance. First, meta-heuristics are fairly simple. They have been mostly inspired by very simple concepts. The inspirations are typically related to physical phenomena, animals’ behaviors, or evolutionary concepts. The simplicity allows computer scientists to simulate different natural concepts, propose new meta-heuristics, hybridize two or more meta-heuristics, or improve the current meta-heuristics. Moreover, the simplicity assists other scientists to learn meta-heuristics quickly and apply them to their problems. Second, flexibility refers to the applicability of meta-heuristics to different problems without any special changes in the structure of the algorithm. Meta-heuristics are readily applicable to different problems since they mostly assume problems as black boxes. In other words, only the input(s) and output(s) of a system are important for a meta-heuristic. So, all a designer needs is to know how to represent his/her problem for metaheuristics. Third, the majority of meta-heuristics have derivation-free mechanisms. In contrast to gradient-based optimization approaches, meta-heuristics optimize problems stochastically. The optimization process starts with random solution(s), and there is no need to calculate the derivative of search spaces to find the optimum. This makes meta-heuristics highly suitable for real problems with expensive or unknown derivative information. Finally, meta-heuristics have superior abilities to avoid local optima compared to conventional optimization techniques. This is due to the stochastic nature of meta-heuristics which allow them to avoid stagnation in local solutions and search the entire search space extensively. The search space of real problems is usually unknown and very complex with a massive number of local optima, so meta-heuristics are good options for optimizing these challenging real problems. The No Free Lunch (NFL) theorem [4] is worth mentioning here. This theorem has logically proved that there is no meta-heuristic best suited for solving all optimization problems. In other words, a particular metaheuristic may show very promising results on a set of problems, but the same algorithm may show poor performance on a different set of problems. Obviously, NFL makes this field of study highly active which results in enhancing current approaches and proposing new meta-heuristics every year. This also motivates our attempts to develop a new meta-heuristic with inspiration from grey wolves. Generally speaking, meta-heuristics can be divided into two main classes: single-solution-based and population-based. In the former class (Simulated Annealing [5] for instance) the search process starts with one candidate solution. This single candidate solution is then improved over the course of iterations. Populationbased meta-heuristics, however, perform the optimization using a set of solutions (population). In this case the search process starts with a random initial population (multiple solutions), and this population is enhanced over the course of iterations. Population-based meta-heuristics have some advantages compared to single solutionbased algorithms: Multiple candidate solutions share information about the search space which results in sudden jumps toward the promising part of search space Multiple candidate solutions assist each other to avoid locally optimal solutions Population-based meta-heuristics generally have greater exploration compared to single solution-based algorithms One of the interesting branches of the population-based meta-heuristics is Swarm Intelligence (SI). The concepts of SI was first proposed in 1993 [6]. According to Bonabeau et al. [1], SI is “The emergent collective intelligence of groups of simple agents”. The inspirations of SI techniques originate mostly from natural colonies, flock, herds, and schools. Some of the most popular SI techniques are ACO [2], PSO [3], and Artificial Bee Colony (ABC) [7]. A comprehensive literature review of the SI algorithms is provided in the next section. Some of the advantages of SI algorithms are: SI algorithms preserve information about the search space over the course of iteration, whereas Evolutionary Algorithms (EA) discard the information of the previous generations SI algorithms often utilize memory to save the best solution obtained so far SI algorithms usually have fewer parameters to adjust SI algorithms have less operators compared to evolutionary approaches (crossover, mutation, elitism, and so on) SI algorithms are easy to implement Regardless of the differences between the meta-heuristics, a common feature is the division of the search process into two phases: exploration and exploitation [8-12]. The exploration phase refers to the process of investigating the promising area(s) of the search space as broadly as possible. An algorithm needs to have stochastic operators to randomly and globally search the search space in order to support this phase. However, exploitation refers to the local search capability around the promising regions obtained in the exploration phase. Finding a proper balance between these two phases is considered a challenging task due to the stochastic nature of meta-heuristics. This work proposes a new SI technique with inspiration from the social hierarchy and hunting behavior of grey wolf packs. The rest of the paper is organized as follows: Section 2 presents a literature review of SI techniques. Section 3 outlines the proposed GWO algorithm. The results and discussion of benchmark functions, semi-real problems, and a real application are presented in Section 4, Section 5, and Section 6, respectively. Finally, Section 7 concludes the work and suggests some directions for future studies |