A cluster-based co-evolutionary optimization method for bilevel multi-objective optimization
Abstract
The research motivation of multi-objective bilevel optimization mainly stems from the need to solve practical problems and improve decision-making efficiency. On the one hand, bilevel optimization helps to solve the complexity and uncertainty in real life, thereby improving decision-making efficiency and robustness. On the other hand, by promoting the development and application of AI technology, bilevel optimization also provides support for sustainable development. Although the application of bilevel optimization has proven to be beneficial in addressing various real-life problems. However, recent studies indicate that achieving both high speed and high-quality optimization through existing algorithms remains challenging. This difficulty arises due to the NP-hard nature of the bilevel optimization problem. The nested structure method, commonly used to tackle this problem, involves each upper level solution independently performing the lower level optimization task. This approach significantly increases the number of evaluations for the lower level. To address this issue, our proposed method leverages the similarity in lower level optimization to group upper level solutions, enabling co-evolution of lower level solutions within the same group. Consequently, this approach substantially reduces the number of evaluations required for lower level solutions. Additionally, our method pairs parents and offspring, the optimized lower level solutions of the parents are utilized to optimize the lower level solutions of the offspring. This approach accelerates the optimization process for the lower level. To validate the effectiveness of our algorithm, we have applied it to a suite of test problems, demonstrating satisfactory performance.
Keywords
1. INTRODUCTION
Today's society faces two major challenges: sustainable development and optimal use of resources. Green computing aims to reduce the environmental impact of computers and information technology, including reducing energy consumption and reducing e-waste. Bilevel multi-objective optimisation is an optimisation framework that considers multiple objective functions at the same time to find a good balance. By incorporating green computing principles, such as minimizing energy consumption and reducing environmental impact, as additional objectives or constraints within the optimization framework, we can effectively prioritize sustainability alongside other performance metrics. The bilevel optimization problem (BLOP) is a nested structural problem, where a lower level optimization problem is embedded within the upper level optimization problem [1]. Bilevel optimization has its roots in game theory [2]. It has gained increased attention due to its potential usage in various fields, such as economic management, resource allocation, energy sustainability, transportation planning, machine learning and medical engineering[3-9]. In some cases, both the upper level and lower level optimization problems may contain multiple conflicting objective functions, leading to multi-objective bilevel optimization problems (MBOPs). Without loss of generality, a MBOP can be expressed as follows:
where
The single-level reduction approach converts a MBOP into a single-level optimization problem and then applies EAs to solve it. For instance, Li et al. used adaptive weighting sum scalarization and Karush-Kuhn-Tucker (KKT) conditions to transform a BMOP to a multi-objective optimization problem (MOP)[12]. Then they proposed an effective smoothing technique to cope with complementarity constraints. Li et al. [14] improved the algorithm presented in [12]. They first used the KKT condition to transform a BMOP into a MOP involving complementarity constraints and then proposed a decomposition-based constrained multi-objective differential EA. Jia et al. converted a MBOP into a MOP based on the primal and dual theory and used multi-objective metaheuristics and constraint processing techniques for optimization[13]. The method of converting a BMOP into a single-level optimization problem is very mature in the field of bilevel optimization, but these methods usually require strict mathematical assumptions, and these assumptions are usually difficult to satisfy in the real world.
The nested method solves the MBOP directly by performing the lower level optimization independently for each upper level solution. For example, Deb et al. used elite nondominated ranking GA or NSGA-Ⅱ for both upper and lower level optimization[10]. This algorithm was subsequently upgraded by Sinha et al. [22]. Their improved version evaluates the upper level only once, which significantly reduces the number of evaluations of the algorithm. Second, it also allows the file members to participate in crossover, which improves the performance of the algorithm. Deb et al. proposed a hybrid EA combined with a local search strategy for optimization[23]. Cai et al. proposed a divide-and-conquer strategy, in which all variables (both upper and lower levels) are divided into multiple mutually exclusive groups and optimized separately[20]. Although these methods are successful, due to the structural characteristics of bilevel optimization, they cannot afford the large number of evaluations required to be consumed by the lower level evaluation when the number of upper level solutions is large. In order to solve the more complex MBOP, many new algorithms are produced, and a genetic algorithm is adopted at both levels. For example, both the upper and lower levels use Particle swarm optimization, and both upper and lower levels use differential evolution (DE).
The surrogate-model-based methods use a surrogate model to approximate the constraints and objective function at the lower level, with the aim of reducing the number of lower level fitness evaluations (FEs). For example, Sinha et al. proposed an approximation-set-mapping approach which used quadratic functions to address the lower level optimization problem[18]. Sinha et al. modeled the lower level decision variables using a value function named m-BLEAQ; they used a quadratic function to estimate unknown lower level decision variables[19]. It is essential to note that the accuracy of the surrogate model has a significant impact on the performance of final solutions.
In the real world, we are often faced with a complex set of problems, which often require multiple goals to be considered at the same time, and each goal may have a mutually constrained relationship. In order to solve this kind of problem, the bilevel optimization algorithm provides an effective solution. The bilevel optimization algorithm decomposes the problem into two levels, the upper optimization part is responsible for generating a set of feasible solutions, and the lower optimization part further finds the lower optimal solutions that meets the constraints according to this set of feasible solutions. This strategy of decomposing the problem enables the bilevel optimization algorithm to comprehensively consider multiple aspects of the problem, so as to obtain more comprehensive and high-quality solutions. Most of the current MBLOPs are from the perspective of game theory, using the concept of objective function values and Pareto optimality. In contrast, Gu et al. are the first to study the convergence characteristics of MBLOPs problems from the perspective of traditional optimization, and consider a minimum and maximum robust version of the multi-objective problem, weighing the optimality of different objectives, and ensuring that each objective obtains a single optimal solution, rather than generating multiple Pareto optimal solutions[24]. Recently, Wang et al. designed a lower level environment selection strategy and an upper level solution regeneration strategy to improve the search efficiency of the algorithm[25]. Inspired by the biological classification of species, Mejía-de-Dios et al. proposed a novel evolutionary framework based on the concept of family[26]. They adopt the concept of science in biology to promote the diversity of solutions.
However, the implementation and application of the bilevel optimization algorithm still face many challenges. First, the complexity of bilevel optimization problems tends to be high, making the solution process very difficult. Secondly, the existing bilevel optimization algorithms are often difficult to balance the relationship between solution speed and solution quality. In order to solve these problems, we need to research and develop new bilevel optimization algorithms to improve the speed and quality of solutions.
In bilevel optimization, the lower level optimization tasks are usually relatively similar for multiple close upper level solutions. Inspired by this, based on the above observations, we design a new algorithm called CCBMO which utilizes match and cluster. To our knowledge, this is the first time that bilevel optimization has been performed using the similarity of parent and child lower level optimization tasks. The main contributions of this article are summarized as follows:
1. Initially, in the upper level decision space, we match a parent to each offspring based on the Euclidean distance. This makes it easier to select a partial solution from the parent's lower solution as the initial solution when performing a lower optimization for a child, greatly reducing the number of lower level optimizations required.
2. Then, the k-means is used to cluster the upper level offspring into
3. Following this, environment selection is performed based on upper level objective values, and selected upper level solutions are chosen for iteration.
The remainder of this paper is structured as follows: Section Ⅱ presents the proposed bilevel optimization algorithm. Section Ⅲ provides a comprehensive experimental study, analysing and discussing the results. Finally, Section Ⅳ summarizes the findings and the potential for further research.
2. THE PROPOSED ALGORITHM
2.1. General framework of proposed algorithm
The framework of the CCBMO algorithm is established by algorithm 1. Initially, a set of
Algorithm 1: Proposed algorithm
Algorithm 2: Lower level optimization
During the main loop, we execute the DE on
Figure 1. We provide an example of a block plot for matching and clustering in Figure 1A and B. (A) Clustering of
Then
In order to select the better solution for each iteration, the selection operator of NSGA-Ⅱ is applied to
1.
2.
3.
Then, the best half of the upper level solutions are added to
2.2. Lower level optimization
algorithm 2 provides the process of lower level optimization. First, an empty set
Figure 2. Comparison of the average FEs statistical results of the three algorithms on TP1-TP2 and DS1-DS5. FEs: Fitness evaluations.
After removing the repeat values in
2.3. Termination criterion
Termination conditions for both upper and lower layers are based on their hypervolume (HV). The rate of convergence of HV for both the upper and lower levels is called the H-metric and is calculated as follows:
where
When the number of iterations exceeds
3. EXPERIMENTAL STUDIES
In this section, the experimental study for investigating the performance of the proposed algorithm is presented. First, we introduce the test problems and parameter settings. Then we briefly describe the comparison algorithm. Afterward, we introduce the performance metrics used to evaluate the algorithm's performance. Finally, we give the experimental results and analyze the results.
3.1. Test problems
Two benchmark test sets, TP and DS [23], were selected for the experimental studies. In the TP test suite, we chose two test problems, TP1 and TP2, and the DS test suite includes five test problems (denoted as DS1-DS5).
The detailed settings for the TP test set and DS test set are provided in Table 1.
Settings of the test problems
Test problem | K | L | ||
TP1 | 20 | 20 | ||
TP2 | 20 | 20 | ||
DS1 | 20 | 20 | 5 | 5 |
DS2 | 20 | 20 | 5 | 5 |
DS3 | 20 | 20 | 5 | 5 |
DS4 | 5 | 40 | 1 | 9 |
DS5 | 5 | 40 | 1 | 9 |
3.2. Compared algorithms and parameter settings
In order to evaluate the performance of the algorithms, two algorithms were chosen for comparison: BLMOCC [20], MOBEA-DPL [27].
(1) BLMOCC: it is an algorithm that uses a knowledge-based variable decomposition strategy to solve a MBOP. The knowledge-based variable decomposition strategy is used throughout the optimization process, and the variables are divided into three groups according to the correlation between the variables in the two levels. Different optimization methods are used independently for different groups.
(2) MOBEA-DPL: it is an algorithm developed on the framework of the nested bilevel multi-objective optimization algorithm. It uses a dual-population optimization strategy to improve the solution of the lower level optimization problem. The first group is used to store nondominated solutions in the lower level, and the second group is used to store upper level solutions that are not dominated by solutions in the first group. Additionally, to increase the effectiveness of the search, the offspring of the upper level solutions are selected from the neighborhood of the best solutions in the existing solutions.
3.3. Performance metrics
We limit the maximum FEs of the upper and lower levels to facilitate comparing the number of true evaluations of different algorithms. The maximum number of evaluations for the upper and lower levels is 50, 000 and 1, 000, 000, respectively.
Without loss of generality, we consider using the inverted generational distance (IGD) [28] and hypervolume (HV) [29] to measure the performance of the algorithms. Both metrics can measure the diversity and convergence of the obtained solutions. A smaller IGD value means better algorithm performance, while the opposite is true for HV values, where a larger HV value means better algorithm performance.
3.4. Comparison between CCBMO and other algorithms on each index
In this case, we measure whether the algorithm can converge well with a small number of FEs. We recorded the average of FEs (including the upper level, lower level, and the sum of both levels), the IGD, and HV for all test problems. Specifically, each algorithm ran independently 21 times.
From Table 2, it can be seen that CCBMO has better results than BLMOCC and MOBEA-DPL in terms of lower level average FEs on TP1-TP2, DS1, and DS3-DS5, while slightly worse than BLMOCC in terms of upper level average FEs. This is because in lower level optimization, we introduce the idea of co-optimization, which optimizes similar lower level optimization problems at the same time, thus greatly reducing the number of optimizations and evaluations at the lower level level. However, as shown in Figure 3, due to the obvious advantage of CCBMO in the average FEs of the lower level, the sum of the average FEs of the upper and lower levels shows the best results, except that MOBEA-DPL outperformed CCBMO and BLMOCC on DS2. Obviously, this is all due to the improvement of the lower level optimization strategy.
Performance comparison between BLMOCC, MOBEA-DPL and CCBMO regarding the average values of FEs on TP and DS
Problem | BLMOCC | MOBEA-DPL | CCBMO | |||||||
FEs: Fitness evaluations. | ||||||||||
TP1 | 20 + 20 | 2, 536 | 320, 281 | 322, 818 | 53, 499 | 768, 080 | 821, 580 | 40, 615 | 250, 271 | 290, 885 |
TP2 | 20 + 20 | 38, 736 | 1, 128, 305 | 1, 167, 041 | 53, 499 | 768, 080 | 821, 580 | 36, 907 | 230, 748 | 267, 654 |
DS1 (5 + 5) | 20 + 20 | 30, 553 | 561, 769 | 592, 322 | 50, 913 | 634, 383 | 685, 295 | 36, 763 | 278, 767 | 315, 530 |
DS2 (5 + 5) | 20 + 20 | 36, 466 | 650, 363 | 686, 829 | 55, 212 | 297, 359 | 352, 571 | 66, 986 | 291, 608 | 358, 594 |
DS3 (5 + 5) | 20 + 20 | 4, 121 | 657, 513 | 661, 633 | 65, 183 | 1, 243, 679 | 1, 308, 862 | 33, 257 | 298, 037 | 331, 293 |
DS4 (1 + 9) | 5 + 40 | 19, 492 | 490, 870 | 510, 362 | 460, 000 | 2, 070, 000 | 2, 530, 000 | 23, 391 | 99, 774 | 123, 165 |
DS5 (1 + 9) | 5 + 40 | 14, 339 | 362, 115 | 376, 454 | 352, 000 | 1, 590, 000 | 1, 942, 000 | 56, 271 | 99, 959 | 156, 230 |
Figure 3. Comparison of the statistical results of the average IGD values of the three algorithms on TP1-TP2 and DS1-DS5. IGD: Inverted generational distance
Tables 3 and 4 provide the comparative results of three algorithms on seven test problems in terms of IGD and HV. Figures 3 and 4 clearly demonstrate that CCBMO outperformed both BLMOCC and MOBEA-DPL. Next we analyze these experimental results in detail.
Performance comparison between BLMOCC, MOBEA-DPL and CCBMO regarding the IGD values on TP1-TP2 and DS1-DS5
Problem | BLMOCC | MOBEA-DPL | CCBMO | |
IGD: Inverted generational distance. | ||||
TP1 | 20 + 20 | 0.1426 | 0.0126 | 0.0128 |
TP2 | 20+20 | 0.0310 | 0.0235 | 0.0178 |
DS1 (5 + 5) | 20 + 20 | 14.4842 | 0.1123 | 1.0858 |
DS2 (5 + 5) | 20 + 20 | 25.5732 | 4.3112 | 0.1854 |
DS3 (5 + 5) | 20 + 20 | 3.9562 | 0.2058 | 1.9772 |
DS4 (1 + 9) | 5 + 40 | 0.0768 | 0.1680 | 0.0607 |
DS5 (1 + 9) | 5 + 40 | 0.3582 | 0.1820 | 0.0326 |
Performance comparison between BLMOCC, MOBEA-DPL and CCBMO regarding the HV values on TP1-TP2 and DS1-DS5
Problem | BLMOCC | MOBEA-DPL | CCBMO | |
HV: Hypervolume. | ||||
TP1 | 20 + 20 | 0.7723 | 0.7026 | 0.6856 |
TP2 | 20 + 20 | 0.4029 | 0.4074 | 0.4277 |
DS1 (5 + 5) | 20 + 20 | 0.0000 | 0.3092 | 0.2813 |
DS2 (5 + 5) | 20 + 20 | 0.0000 | 0.0028 | 0.2443 |
DS3 (5 + 5) | 20 + 20 | 0.0000 | 0.2342 | 0.0000 |
DS4 (1 + 9) | 5 + 40 | 2.3110 | 0.1680 | 2.3495 |
DS5 (1 + 9) | 5 + 40 | 1.1756 | 0.1820 | 2.2105 |
Figure 4. Comparison of the statistical results of the average HV value of the three algorithms on TP1-TP2 and DS1-DS5. HV: Hypervolume.
Both TP1 and TP2 are non-scalability problems. MOBEA-DPL outperformed other algorithms in terms of IGD value for TP1 due to its lower level dual-population strategy, but it inevitably increased the lower level FEs. CCBMO demonstrates a slightly inferior performance in terms of IGD value compared to MOBEA-DPL. Among all the compared algorithms, BLMOCC obtained the best HV value but the worst IGD value. Regarding TP2, CCBMO achieved the most satisfactory IGD and HV, followed by MOBEA-DPL. BLMOCC performed the worst.
DS1 is often used to evaluate algorithm performance and the ability of the algorithm to coordinate the processing of upper level and lower level tasks. Regarding DS1, MOBEA-DPL achieved the best IGD and HV. CCBMO performed slightly inferior to MOBEA-DPL in terms of IGD and HV values but significantly outperformed BLMOCC.
DS2 is used to assess the algorithm's capability to conduct extensive searches before converging on the frontier. For DS2, CCBMO significantly outperformed the contrasting algorithms. Compared with BLMOCC and MOBEA-DPL, CCBMO could reduce the IGD value by one to two orders of magnitude, respectively, and increase the HV value by two orders of magnitude. BLMOCC and MOBEA-DPL had difficulty achieving satisfactory results in the DS2 test problem.
DS3 involves discrete variables, and as the number of variables increases, the problem becomes increasingly challenging, making it difficult for traditional algorithms to address. MOBEA-DPL achieved the best IGD and HV on DS3. In contrast, CCBMO performed worse than MOBEA-DPL due to its utilization of a traditional nesting method for addressing this problem. BLMOCC had the worst IGD and HV.
Both DS4 and DS5 evaluate the algorithm's capability to search for the appropriate lower level frontier that corresponds to the upper level frontier. Additionally, DS4 necessitates identifying a specific point in the lower level solution that corresponds to the upper level solution. In the case of DS4, CCBMO achieved superior IGD and HV values, which are one and two orders of magnitude better than the corresponding IGD values of MOBEA-DPL and BLMOCC, respectively, and one order of magnitude better than the HV value of MOBEA-DPL. In the case of DS5, CCBMO had significant superiority over the comparison algorithms. In terms of IGD value, CCBMO outperformed both comparison algorithms by two orders of magnitude. The HV value of CCBMO is an order of magnitude better than that of MOBEA-DPL. Therefore, these experiments demonstrate the effectiveness of CCBMO in solving BMOPs.
3.5. Comparison of CCBMO with its variants
To illustrate the effectiveness of the mechanism in CCBMO, we compared two variants of CCBMO, CCBMO-M and CCBMO-C, where CCBMO-M represents the variant that removes the k-means clustering in CCBMO, and CCBMO-C represents the variant that removes the matching between the upper level offspring and parents in CCBMO. Tables 5-7 present the results of CCBMO, CCBMO-M and CCBMO-C regarding the average number of FEs, IGD and HV over 21 runs on TP1-TP2 and DS1-DS5, respectively.
FEs statistics of CCBMO and its variants on TP1-TP2 and DS1-DS5
Problem | CCBMO | CCBMO-M | CCBMO-C | |
FEs: Fitness evaluations. | ||||
TP1 | 20 + 20 | 290, 885 | 292, 893 | 334, 300 |
TP2 | 20 + 20 | 267, 654 | 290, 721 | 358, 603 |
DS1 (5 + 5) | 20 + 20 | 315, 530 | 312, 516 | 363, 501 |
DS2 (5 + 5) | 20 + 20 | 358, 594 | 367, 339 | 379, 283 |
DS3 (5 + 5) | 20 + 20 | 331, 293 | 319, 998 | 406, 935 |
DS4 (1 + 9) | 5 + 40 | 123, 165 | 126, 272 | 147, 857 |
DS5 (1 + 9) | 5 + 40 | 156, 230 | 161, 888 | 119, 948 |
IGD statistics of CCBMO and its variants on TP1-TP2 and DS1-DS5
Problem | CCBMO | CCBMO-M | CCBMO-C | |
IGD: Inverted generational distance. | ||||
TP1 | 20 + 20 | 0.0128 | 0.0156 | 0.0192 |
TP2 | 20 + 20 | 0.0178 | 0.1199 | 0.6835 |
DS1 (5 + 5) | 20 + 20 | 1.0858 | 1.0070 | 3.7125 |
DS2 (5 + 5) | 20 + 20 | 0.1854 | 0.2088 | 0.3523 |
DS3 (5 + 5) | 20 + 20 | 1.9772 | 2.1425 | 2.9403 |
DS4 (1 + 9) | 5 + 40 | 0.0607 | 0.0684 | 0.0635 |
DS5 (1 + 9) | 5 + 40 | 0.0326 | 0.0360 | 0.0571 |
HV statistics of CCBMO and its variants on TP1-TP2 and DS1-DS5.
Problem | CCBMO | CCBMO-M | CCBMO-C | |
HV: Hypervolume. | ||||
TP1 | 20 + 20 | 0.6856 | 0.6298 | 0.6102 |
TP2 | 20 + 20 | 0.4277 | 0.4401 | 0.0239 |
DS1 (5 + 5) | 20+20 | 0.2813 | 0.3559 | 0.0000 |
DS2 (5 + 5) | 20+20 | 0.2443 | 0.1919 | 0.0759 |
DS3 (5 + 5) | 20+20 | 0.0000 | 0.0000 | 0.0000 |
DS4 (1 + 9) | 5+40 | 2.3495 | 2.2879 | 2.3383 |
DS5 (1 + 9) | 5+40 | 2.2105 | 2.2082 | 2.1234 |
Table 5 gives a comparison of CCBMO and its variants in terms of average FEs. CCBMO consumed the least number of evaluations in solving TP1-TP2, DS2 and DS4. CCBMO-M performed slightly worse than CCBMO, consuming the least number of evaluations on solving DS1 and DS3. CCBMO-C only consumed the least number of evaluations when solving DS5.
Tables 6 and 7 compare the IGD and HV values of CCBMO and its variants (CCBMO-M and CCBMO-C), respectively. In terms of TP1 and DS2-DS5, CCBMO outperformed CCBMO-M and CCBMO-C in both IGD and HV values. For TP2, CCBMO obtained the best IGD value, but the HV value was slightly worse than CCBMO-M. CCBMO-M had the best IGD and HV values when solving DS1. On the other hand, CCBMO-C performed the worst, not obtaining the best IGD and HV values for any problem. Therefore, our experiments demonstrate the effectiveness of matching and clustering incorporated in CCBMO.
4. CONCLUSION
Our work shows the existing bilevel multi-objective optimization algorithms ignore the similarity between the lower level optimization tasks when solving the problem, and thus consume a large number of FEs when performing the lower level optimization. Based on these findings, we have proposed a new optimization algorithm framework (called CCBMO), which utilizes the similarity of the lower level optimization tasks of the parents and offspring to accelerate the solution of the lower level optimization tasks of the offspring. CCBMO mainly includes two stages: a matching stage and a clustering stage. In the matching stage, upper level solutions of the parents and offspring are matched based on Euclidean distance. Each offspring selects the closest parent and inherits its corresponding lower level solutions. The clustering stage subsequently decomposes the offspring into
Although CCBMO performed well on most problems, we found that CCBMO did not converge well when solving test problems with a large number of decision variables. Therefore, further efforts are needed to develop an efficient selection strategy to address larger-scale BMOPs.
DECLARATIONS
Acknowledgments
Thanks to Yaochu Jin, Xingyi Zhang, Ran Cheng, Ye Tian, Xilu Wang, Dan Guo, Dimo Brockhoff, Handing Wang Qingfu Zhang and Chaoli Sun for providing their source code.
Authors' contributions
Proposal and design of research propositions, drafting or final revision of the paper: Hu WY, Huang PQ
Perform data collection: Hu WY
Technical and material support provided: Li F, Ge EQ
Availability of data and materials
Not applicable.
Financial support and sponsorship
The authors is supported by the National Natural Science Foundation of China under 61903003 and supported by Anhui Provincial Natural Science Foundation (Grant 2308085MF199 and 2008085QE227) and Scientific Research Projects in Colleges and Universities of Anhui Province (Grant 2023AH051124) and Supported by the Open Project of Anhui Province Engineering Laboratory of Intelligent Demolition Equipment (Grant APELIDE2022A007) and Supported by the Open Project of Anhui Province Key Laboratory of Special and Heavy Load Robot (Grant TZJQR001-2021).
Conflicts of interest
All authors declare that they are bound by confidentiality agreements that prevent them from disclosing their conflicts of interest in this work.
Ethical approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Copyright
© The Author(s) 2024.
REFERENCES
1. Dempe S, Dinh N, Dutta J, Pandit T. Simple bilevel programming and extensions. Math Program 2021;188:227-53.
3. Fortuny-Amat J, McCarl B. A representation and economic interpretation of a two-level programming problem. J Oper Res Soc 1981;32:783-92.
4. Dempe S, Kalashnikov V, Pérez-Valdés GA, Kalashnykova N. Bilevel programming problems Heidelberg: Springer Berlin; 2015.
5. Wogrin S, Pineda S, Tejada-Arango DA. Applications of bilevel optimization in energy and electricity markets. In: Dempe S, Zemkoho A, editors. Bilevel optimization. Cham: Springer; 2020. pp. 139-68.
6. Yin Y. Multiobjective bilevel optimization for transportation planning and management problems. J Adv Transp 2002;36:93-105.
7. Liu R, Gao J, Zhang J, Meng D, Lin Z. Investigating bi-level optimization for learning and vision from a unified perspective: a survey and beyond. IEEE Trans Pattern Anal Mach Intell 2022;44:10045-67.
9. Halter W, Mostaghim S. Bilevel optimization of multi-component chemical systems using particle swarm optimization. In: 2006 IEEE International Conference on Evolutionary Computation; 2006 Jul 16-21; Vancouver, Canada. IEEE; 2006. pp. 1240-47.
10. Deb K, Sinha A. Solving bilevel multi-objective optimization problems using evolutionary algorithms. In: Ehrgott M, Fonseca CM, Gandibleux X, Hao JK, Sevaux M, editors. Lecture notes in computer Science. Heidelberg: Springer Berlin; 2009. pp. 110-24.
11. Jia L, Wang Y. A genetic algorithm for multiobjective bilevel convex optimization problems. In: 2009 International Conference on Computational Intelligence and Security; 2009 Dec 11-14; Beijing, China. IEEE; 2009. pp. 98-102.
12. Li H, Zhang L. An efficient solution strategy for bilevel multiobjective optimization problems using multiobjective evolutionary algorithm. Soft Comput 2021;25:8241-61.
13. Jia L, Wang Y. Genetic algorithm based on primal and dual theory for solving multiobjective bilevel linear programming. In: 2011 IEEE Congress of Evolutionary Computation (CEC); 2011 Jun 5-8; New Orleans, LA, USA. IEEE; 2011. pp. 558-65.
14. Li H, Zhang Q, Chen Q, Zhang L, Jiao YC. Multiobjective differential evolution algorithm based on decomposition for a type of multiobjective bilevel programming problems. Knowl Based Syst 2016;107:271-88.
15. Pieume CO, Marcotte P, Fotso LP, Siarry P. Generating efficient solutions in bilevel multi-objective programming problems. Am J Oper Res 2013;3:289-98.
16. He Q, Lv Y. Particle swarm optimization based on smoothing approach for solving a class of bi-level multiobjective programming problem. Cybern Inf Technol 2017;17:59-74.
17. Alves MJ, Dempe S, Júdice JJ. Computing the Pareto frontier of a bi-objective bi-level linear problem using a multiobjective mixed-integer programming algorithm. Optimization 2010;61:335-58.
18. Sinha A, Malo P, Deb K. Approximated set-valued mapping approach for handling multiobjective bilevel problems. Comput Oper Res 2017;77:194-209.
19. Sinha A, Malo P, Deb K. Towards understanding bilevel multi-objective optimization with deterministic lower level decisions. In: Gaspar-Cunha A, Henggeler Antunes C, Coello C, editors. Lecture notes in computer science. Cham: Springer; 2015. pp. 426-43.
20. Cai X, Sun Q, Li Z, et al. Cooperative coevolution with knowledge-based dynamic variable decomposition for bilevel multiobjective optimization. IEEE Trans Evol Comput 2022;26:1553-65.
21. Islam MM, Singh HK, Ray T. A nested differential evolution based algorithm for solving multi-objective bilevel optimization problems. In: Ray T, Sarker R, Li X, editors. Lecture notes in computer science. Cham: Springer; 2016. pp. 101-12.
22. Sinha A, Deb K. Towards understanding evolutionary bilevel multi-objective optimization algorithm. IFAC Proc Vol 2009;42:338-43.
23. Deb K, Sinha A. An efficient and accurate solution methodology for bilevel multi-objective programming problems using a hybrid evolutionary-local-search algorithm. Evol Comput 2010;18:403-49.
24. Gu A, Lu S, Ram P, Weng TW. Min-max multi-objective bilevel optimization with applications in robust machine learning. arXiv. [Preprint. ] Mar 7, 2023[accessed 2023 Dec 21]. Available from: https://arxiv.org/abs/2203.01924.
25. Wang W, Liu HL. A multi-objective bilevel optimization evolutionary algorithm with local search. In: 2021 17th International Conference on Computational Intelligence and Security (CIS); 2021 Nov 19-22; Chengdu, China. IEEE; 2021. pp. 408-12.
26. Mejía-de-Dios JA, Rodríguez-Molina A, Mezura-Montes E. A novel evolutionary framework based on a family concept for solving multi-objective bilevel optimization problems. In: Proceedings of the Genetic and Evolutionary Computation Conference Companion. New York, NY, USA: Association for Computing Machinery; 2022. pp. 348–51.
27. Wang W, Liu HL, Shi H. A multi-objective bilevel optimisation evolutionary algorithm with dual populations lower-level search. Conn Sci 2022;34:1556-81.
28. Bosman PAN, Thierens D. The balance between proximity and diversity in multiobjective evolutionary algorithms. IEEE Trans Evol Comput 2003;7:174-88.
Cite This Article
How to Cite
Hu, W. Y.; Li F.; Huang P. Q.; Ge E. Q. A cluster-based co-evolutionary optimization method for bilevel multi-objective optimization. . J. Smart. Environ. Green. Comput. 2024, 4, 4-21. http://dx.doi.org/10.20517/jsegc.2023.18
Download Citation
Export Citation File:
Type of Import
Tips on Downloading Citation
Citation Manager File Format
Type of Import
Direct Import: When the Direct Import option is selected (the default state), a dialogue box will give you the option to Save or Open the downloaded citation data. Choosing Open will either launch your citation manager or give you a choice of applications with which to use the metadata. The Save option saves the file locally for later use.
Indirect Import: When the Indirect Import option is selected, the metadata is displayed and may be copied and pasted as needed.
Comments
Comments must be written in English. Spam, offensive content, impersonation, and private information will not be permitted. If any comment is reported and identified as inappropriate content by OAE staff, the comment will be removed without notice. If you have any queries or need any help, please contact us at support@oaepublish.com.