This paper presents an extensive analysis of the sensitivity of multi-objective algorithm parameters and objective function scaling tested on a large number of parameter setting combinations for a water distribution system optimisation problem. The optimisation model comprises two operational objectives minimised concurrently, the pump energy costs and deviations of constituent concentrations as a water quality measure. This optimisation model is applied to a regional non-drinking water distribution system, and solved using the optimisation software GANetXL incorporating the NSGA-II linked with the network analysis software EPANet. The sensitivity analysis employs a set of performance metrics, which were designed to capture the overall quality of the computed Pareto fronts. The performance and sensitivity of NSGA-II parameters using those metrics is evaluated. The results demonstrate that NSGA-II is sensitive to different parameter settings, and unlike in the single-objective problems, a range of parameter setting combinations appears to be required to reach a Pareto front of optimal solutions. Additionally, inadequately scaled objective functions cause the NSGA-II bias towards the second objective. Lastly, the methodology for performance and sensitivity analysis may be used for calibration of algorithm parameters.

## NOTATION

*a*_{i}area determined by the solutions in PF

_{known}and the reference point*b*_{mi}binary variable describing the status of pump

*m*during a time interval*i**c*constituent (water quality parameter)

*C*number of constituents (water quality parameters)

*c**c*_{cdj}actual (modelled) concentration of a constituent

*c*at the customer node*d*during a time interval*j*maximum allowed concentration of a constituent

*c*at the customer node*d*minimum required concentration of a constituent

*c*at the customer node*d*- CRS
crossover operator

*d*customer demand node

*D*number of customer demand nodes

*d*maximum allowed volume deficit in storage tank

*s*at the end of the simulation period*T**E*metric error ratio

*EC*_{i}electricity tariff during a time interval

*i**EX*metric extent

*FI*(*b*)economic objective function as pump energy costs, over the simulation period

*T**FII*(*b*)water quality objective function as squared deviations of actual constituent concentrations from required values, over the simulation period

*T**GD*metric generational distance

- GEN
number of generations

*H*_{di}pressure at customer demand node

*d*during a time interval*i*minimum required pressure at customer demand node

*d**H*_{mi}total dynamic head supplied by pump

*m*during a time interval*i**i*time interval for hydraulic simulation

*IE*metric

*ɛ*-indicator*j*time interval for water quality simulation

*k*unit conversion coefficient

*L*number of equal time intervals

*i*for a hydraulic simulation within the simulation period*T**LL*number of equal time intervals

*j*for water quality simulation within the simulation period*T**m*pump

*M*number of pumps

*M*_{3}metric extent (Zitzler 1999; Zitzler

*et al.*2000)- MUT
mutation operator

*n*non-demand node

*N*total number of nodes within the network

*NN*metric non-dominated number

*NN’*number of non-dominated solutions in PF

_{known}*ONVG*metric overall non-dominated vector generation

*p*pipe

*P*pump station

- PF
_{know}known (i.e. computed) Pareto front

- PF
_{true}true Pareto front

*POP,*POPpopulation size

*Q*_{mi}flow through pump

*m*during a time interval*i**r*reservoir

*s*storage tank

*S*number of storage tanks

*s*- S1P
simple one point (crossover)

*SC*metric spacing

- SEG
sensitivity group

- SPs
sensitivity points

- SXP
simple multi point (crossover)

*SM*S-metric

*TN*metric true number

*UN*metric unique non-dominated number

*V*_{s}volume of storage tank

*s*constrained by minimum and maximum water levels*y*_{si}water level in storage tank

*s*during a time interval*i*minimum water level in storage tank

*s*maximum water level in storage tank

*s*value of the

*i*th objective in PF_{known}value of the

*i*th objective in PF_{true}*δ*_{i}Euclidean distance between

*i*th solution in PF_{known}and the nearest solution in PF_{true}*Δt*_{i}length of a time interval

*i**ΔV*_{sT}volume deficit (difference between the initial and final volume) in storage tank

*s*at the end of the simulation period*T**ɛ*_{i}Euclidean distance between

*i*th solution and its closest neighbour in PF_{known}mean of all

*ɛ*_{i}*ɛ*_{outer}Euclidean distance between the two outer solutions in PF

_{known}Euclidean distance between the two outer solutions in the PF

_{true}*η*_{i}solution in PF

_{known}*η*_{m}efficiency of pump

*m*

## INTRODUCTION

Optimisation of water distribution systems (WDSs), which generally covers optimum system design, rehabilitation and operation, has been a subject of intensive research over the past several decades. Although these problems inherently involve multiple objectives, they were initially approached as single-objective due to the computational complexity involved. Over the past two decades, nonetheless, multi-objective approaches have been increasingly applied to analyse trade-offs between competing objectives; for example system and operational costs, service quality, system reliability and water quality. Regarding optimum design and rehabilitation, the minimisation of the network costs is often defined as the first objective, with other competing objectives being the maximisation of network reliability (Prasad & Park 2004; Alvisi & Franchini 2006; Farmani *et al.* 2006, Farmani *et al.* 2005b; Giustolisi & Berardi 2009; Raad *et al.* 2010), network robustness (Kapelan *et al.* 2005; Babayan *et al.* 2007) or network performance (Halhal *et al.* 1999; Walters *et al.* 1999; Vamvakeridou-Lyroudia *et al.* 2005). Additionally, network design and/or rehabilitation costs were traded off against service quality measures such as pressure head deficit (Farmani *et al.* 2005a; Atiquzzaman *et al.* 2006; Keedwell & Khu 2006; Jin *et al.* 2008). Optimum operation of WDSs has been mainly concerned with the minimisation of energy costs for pumping as the first objective and pump maintenance as the second objective (Savic *et al.* 1997; Sotelo & Baran 2001; Kelner & Leonard 2003; Lopez-Ibanez *et al.* 2005). Most recently, tradeoffs between pump energy costs and greenhouse gas emissions (Wu *et al.* 2010) or water quality measures (Kurek & Ostfeld 2013, 2014) have also been investigated.

Scaling of the objective and constraint functions is important, because it may have a significant influence on the performance of optimisation algorithms. It represents down-scaling or up-scaling of these functions, so that they cover a similar range of numerical values. However, the majority of the above literature did not consider scaling of the objective functions. Inadequate scaling of the objective functions may lead, according to Bhattacharjya (2007), to the algorithm being biased towards one objective, specifically an objective with larger objective function values. It may affect spacing between the solutions with concentrations of solutions either increasing or decreasing along the Pareto front and may cause discontinuity of the Pareto front (Bhattacharjya 2007). It may also impact the location of the Pareto front with solutions being away from the true Pareto front (Bhattacharjya 2007). More detailed information on scaling of optimisation problems can be found in Gill *et al.* (1981). Work presented within this paper addresses those scaling issues by comparing a broad range of results with non-scaled and scaled objective functions for a multi-objective problem of the optimal operation of WDSs.

To solve multi-objective optimisation problems in WDSs, multiple objective evolutionary algorithms (MOEAs) have been mostly applied. Those MOEAs include non-dominated sorting genetic algorithm (NSGA) (Srinivas & Deb 1994), NSGA-II (Deb *et al.* 2002), strength Pareto evolutionary algorithm (SPEA) (Zitzler & Thiele 1999), SPEA2 (Zitzler *et al.* 2001) and others. These algorithms use certain parameters being population size, probability of mutation and crossover, which need to be set up (i.e. calibrated) for each individual problem, so the algorithm delivers optimal solution (Gibbs *et al.* 2010; Younis & Dandy 2012; Marchi *et al.* 2014). Nevertheless, very limited attention has been dedicated to date to analyse sensitivity of algorithm parameters and their performance in multi-objective optimisation of WDSs. In the majority of literature reviewed, the selection of parameter settings is either not discussed (Halhal *et al.* 1999; Walters *et al.* 1999; Kelner & Leonard 2003; Farmani *et al.* 2006; Babayan *et al.* 2007; Raad *et al.* 2010) or is based on literature values (Farmani *et al.* 2005a; Alvisi & Franchini 2006; Alfonso *et al.* 2010) with no further rationale provided. There are also cases where the same algorithm with the same parameter settings is applied to different test networks (Farmani *et al.* 2005a; Keedwell & Khu 2006), which contradicts the importance of calibrating the algorithm parameter values for each individual problem as highlighted by Gibbs *et al.* (2010). Although some authors tested a range of parameter settings (Kapelan *et al.* 2005; Vamvakeridou-Lyroudia *et al.* 2005; Kurek & Ostfeld 2014), they did not present such analysis as part of their research. This current work explores the sensitivity of multi-objective algorithm parameters, evaluates the performance of a wide range of parameter settings and establishes the influence of these settings on solutions for the optimal operation of WDSs including two objectives. It utilises a set of performance metrics to do so.

Performance metrics were proposed in Van Veldhuizen (1999) and Zitzler (1999) and are normally used to compare performances of different multi-objective algorithms (Van Veldhuizen & Lamont 2000b; Zitzler *et al.* 2000; Kollat & Reed 2006; Kollat *et al.* 2008). According to Zitzler *et al.* (2003), the comparison of solutions in multi-objective optimisation is substantially more complex than in single-objective, because there is no single performance metric which is both compatible and complete. Zitzler *et al.* (2000) suggest that such a comparison evaluates (i) distance of the obtained Pareto front from the Pareto front of optimal solutions (i.e. true Pareto front), (ii) uniformity of distribution of the solutions in the Pareto front and (iii) the extent of the obtained Pareto front to ensure that a wide range of objective values is covered. In WDS optimisation, performance metrics were applied in optimal system design (Keedwell & Khu 2006; Raad *et al.* 2010; Kanta *et al.* 2012) and operation (Baran *et al.* 2005) to compare performances of multiple MOEAs. To the authors' knowledge, there are no WDS-focused studies which have analysed the sensitivity of algorithm parameters within one MOEA. Early studies were applied to general test cases (Hadka & Reed 2012) or general water resources applications (Reed *et al.* 2012), and used performance metrics and Sobol's global variance decomposition to evaluate parameter sensitivity.

The purpose of this current work is to comprehensively explore the sensitivity of multi-objective algorithm parameters and objective function scaling. This work is believed to be the first dealing with the performance and sensitivity of algorithm parameters within one MOEA in WDS optimisation. Algorithm parameters considered include population size, number of generations, probability of crossover (subsequently referred to as crossover) and probability of mutation (subsequently referred to as mutation). The test network adopted from the literature (Ostfeld & Salomons 2004; Ostfeld *et al.* 2011) represents a regional non-drinking WDS, where two operational objectives, economic and water quality, are formulated to be optimised. It was identified in both urban drinking WDSs (Kurek & Ostfeld 2014) and regional non-drinking WDSs (Mala-Jetmarova *et al.* 2014) that a relationship between the economic objective (i.e. pump energy costs) and water quality objective is conflicting. The reason for this relationship is that, with more relaxed water quality requirements, increased flexibility exists for pumps to operate more effectively over a broader objective space (Mala-Jetmarova *et al.* 2014).

## OPTIMISATION MODEL

### Objective functions

#### Economic objective

*et al.*2008). Explicit pump schedules are used, which specify the time when a pump operates. For this purpose, simulation period

*T*is divided into equal time intervals

*i*, in which the pump

*m*adopts a binary variable

*b*of either 0 or 1 to describe its status as being off or on, respectively. The economic objective function for the pump scheduling problem is written as: where

_{mi}*M*(

*-*) is the number of pumps,

*m=1, …, M; L*(

*-*) is the number of equal time intervals

*i*for a hydraulic simulation within the simulation period

*T, i=1, …, L*;

*EC*($/kWh) is the electricity tariff during a time interval

_{i}*i; k*(

*-*) is a unit conversion coefficient;

*H*(

_{mi}*m*) is the total dynamic head supplied by pump

*m*during a time interval

*i*;

*Q*(

_{mi}*L/s*) is the flow through pump

*m*during a time interval

*i*;

*η*(

_{m}*-*) is the efficiency of pump

*m; Δt*(

_{i}*s*) is the length of a time interval

*i; b*(

_{mi}*-*)

*, b*is the binary variable describing the status of pump

_{mi}=0, 1*m*as being off or on, respectively, during a time interval

*i*.

#### Water quality objective

where *C* (*-*) is the number of constituents (water quality parameters) *c, c=1, …, C*; *D* (*-*) is the number of customer demand nodes *d, d=1, …, D; LL* (*-*) is the number of equal time intervals *j* for water quality simulation within the simulation period *T, j=1, …, LL*; and (mg/L) are the minimum required and maximum allowed concentrations of a constituent *c* at the customer node *d,* respectively; *c _{cdj}* (

*mg/L*) is the actual (modelled) concentration of a constituent

*c*at the customer node

*d*during a time interval

*j*.

### Constraints and decision variables

*H*(

_{di}*m*) is pressure at demand node

*d*during a time interval

*i*; (

*m*) is the minimum required pressure at demand node

*d*;

*y*(

_{si}*m*) is the water level in storage tank

*s*during a time interval

*i*; (

*m*) and (

*m*) is the minimum and maximum water levels in storage tank

*s,*respectively;

*ΔV*(

_{sT}*m*) is volume deficit (difference between the initial and final volume) in storage tank

^{3}*s*at the end of the simulation period

*T; V*(

_{s}*m*) is volume of storage tank

^{3}*s*constrained by minimum and maximum water levels

*;*(

*%*) is the maximum allowed volume deficit in storage tank

*s*at the end of the simulation period

*T*.

The optimisation problem includes one type of decision variable, which is the binary variable *b _{mi} ɛ {0, 1}* describing the status of pump

*m*(off or on, respectively) during a time interval

*i*.

### Multi-objective optimisation problem

## MODEL APPLICATION

The optimisation model is applied to a network shown in Figure 1, which represents a regional WDS, where non-drinking water is supplied to various customer groups, for example, agricultural, domestic and industrial. Requirements for water quality vary across these customer groups. This network was adapted from Ostfeld & Salomons (2004) with minor modifications from Ostfeld *et al.* (2011). It consists of 11 pipes (*p _{1}–p_{11}*)

*,*six customer demand nodes (

*d*)

_{1}–d_{6}*,*three non-demand nodes (

*n*)

_{1}–n_{3}*,*eight pumps (

*m*) located at three pump stations (

_{1}–m_{8}*P*) and an elevated storage tank (

_{1}–P_{3}*s*)

_{1}*.*The system is supplied from three distinct water sources (reservoirs) (

*r*) of different qualities.

_{1}–r_{3}Water quality is represented by one non-specified conservative (i.e. non-reactive) water quality parameter with constituent concentration (mg/L), which differs from reservoir to reservoir, but is constant over time*.* Each customer group (*d _{1}–d_{6}*) has different water quality requirements characterised by the maximum value of the concentration. Data for the network can be found in Mala-Jetmarova

*et al.*(2013).

The optimisation is undertaken for an extended period simulation (EPS) with the length of the simulation period being 24 hours, and the simulation time step for both hydraulic and water quality analysis being 1 hour. Because the emphasis of this work is not concerned with detailed system operations, but rather sensitivity of the NSGA-II parameters, these settings are deemed acceptable for the purpose of this work.

## SOLUTION SCHEME

The solution methodology integrates a network analysis simulator with an optimisation model (Figure 2). Network analysis software EPANet (Rossman 2000) version 2.0 is used to perform hydraulic and water quality extended period simulations (EPS). EPANet is a widely accepted open-source software package, which is frequently used in conjunction with optimisation research for WDSs.

## GANetXL

Software GANetXL (CWS 2011), available from the Centre for Water Systems (CWS), University of Exeter, UK, is used as the optimisation tool. GANetXL, which was developed by the CWS as a Microsoft Excel add-in, is a generic optimisation engine with spreadsheet based interface for solving both single and multi-objective optimisation problems (Savic *et al.* 2011). The foremost advantage of the GANetXL is that it enables easy integration with the EPANet via visual basic for applications (VBA) code and also enables direct access to evolutionary single and multi-objective algorithms. From the family of MOEAs, GANetXL incorporates the NSGA-II (Deb *et al.* 2002). This work confirmed that GANetXL is well suited for solving multi-objective optimisation problems in WDS operation, which incorporate water quality aspects with conservative constituents.

## NSGA-II

The NSGA-II is a population based multi-objective optimisation algorithm, which was introduced by Deb *et al.* (2000). Since then, this algorithm has been successfully applied to multi-objective optimisation problems of the design and rehabilitation (Prasad & Park 2004; Farmani *et al.* 2005b; Atiquzzaman *et al.* 2006; Jin *et al.* 2008; Kanta *et al.* 2012), and operation (Baran *et al.* 2005) of WDSs. The NSGA-II principally works as follows. An initial population *P _{0}* of random solutions of size

*N*is generated and each solution assigned a fitness (or rank). Applying binary tournament selection, and crossover and mutation operators

*,*the offspring population

*Q*of size

_{0}*N*is created. Populations

*P*and

_{0}*Q*are combined into the population

_{0}*R*=

_{0}*P*∪

_{0}*Q*of the size

_{0}*2N*. Non-dominated sorting is then applied, where the population

*R*is sorted into sets of non-dominated fronts or sets of solutions which are dominated by the same number of Pareto fronts. From these sets

_{0}*F*of the combined population

_{1}, F_{2}, F_{3}, … ,F_{n}*R*the set

_{0},*F*includes the best and

_{1}*F*worst solutions. If the size of the set

_{n}*F*is smaller than

_{1}*N,*all members of this set are selected for the new population

*P*. The remaining places in the population

_{t}*P*are filled in subsequently from sets

_{t}*F*and so on until the size

_{2}, F_{3}*N*is reached, using both the rank and crowding distance. The solutions in the set, which as the last set contributes to the new population

*P*are sorted using the crowded-comparison operator and the best solutions are selected to fill in the last places in

_{t},*P*. The offspring population

_{t}*Q*is then created from

_{t}*P*applying selection, crossover and mutation operators, where selection is based on crowded-comparison operator. The process is repeated by creating the population

_{t}*R*∪

_{t}=P_{t}*Q*of the size

_{t}*2N,*forming non-dominated sets

*F*and creating the new population

_{1}, F_{2}, F_{3}, …, F_{n}*P*and so on, until stopping criteria is met. A detailed description of the NSGA-II can be found in Deb

_{t+1}*et al.*(2000, 2002).

## METHODOLOGY FOR PERFORMANCE AND SENSITIVITY ANALYSIS

The value of this work lies chiefly in the extensive performance and sensitivity analysis of the NSGA-II parameters and objective function scaling, which is tested on a wide range of parameter settings. The NSGA-II parameters considered are: (i) population size, (ii) the number of generations, (iii) crossover and (iv) mutation. A reasonable range for each of these parameters was identified based on the recommended values from within literature and also preliminary testing, and is included in Table 1. Mutation and crossovers can also adopt different types. The type selected for mutation is ‘simple by gene’, which is set as the default within GANetXL. Two types were tested for crossover, being ‘simple one point’ (S1P) and ‘simple multi-point’ (SXP). These mutation and crossover types are described in more detail in CWS (2011).

Parameter . | Parameter setting . | Parameter type . | Comment . |
---|---|---|---|

Population size | 50; 100; 150; 200; 300; 400 | N/A | Values based on preliminary testing |

Number of generations | 500; 1,000; 2,000 | N/A | Values based on preliminary testing |

Mutation operator | 0.002; 0.005*; 0.01; 0.1 | Simple by gene | *0.005 ≈ 1/192, where the value of 192 represents the number of decision variables (8 pumps × 24 1-hour intervals) (Deb et al. 2002) |

Crossover operator | 0.6; 0.7; 0.8; 0.9; 1.0 | Simple one point | Values based on the recommendation of (Goldberg 1989) |

Simple multi point |

Parameter . | Parameter setting . | Parameter type . | Comment . |
---|---|---|---|

Population size | 50; 100; 150; 200; 300; 400 | N/A | Values based on preliminary testing |

Number of generations | 500; 1,000; 2,000 | N/A | Values based on preliminary testing |

Mutation operator | 0.002; 0.005*; 0.01; 0.1 | Simple by gene | *0.005 ≈ 1/192, where the value of 192 represents the number of decision variables (8 pumps × 24 1-hour intervals) (Deb et al. 2002) |

Crossover operator | 0.6; 0.7; 0.8; 0.9; 1.0 | Simple one point | Values based on the recommendation of (Goldberg 1989) |

Simple multi point |

Regarding scaling of the objective function, two alternatives are considered. Firstly, objective functions are not scaled. Secondly, the water quality objective function is linearly scaled down to ensure that the two objective functions have a similar range of numerical values. As an example, the range of the water quality objective as deviations of constituent concentrations is 0–100,000 (mg/L)^{2} and the range of the economic objective as pump energy costs is 150–210 $/day. Hence, the water quality objective is multiplied by 0.01 to obtain a similar order to the economic objective. The first case is referred to subsequently in the text as non-scaled objective functions, and the second case as scaled objective functions.

NSGA-II parameters settings were systematically arranged into 76 parameter setting combinations with either non-scaled or scaled objective functions (Table 2), so the total number of analyses performed was 152 (i.e. 2 × 76). Due to the large number of modelling combinations the optimisation runs were undertaken as a two stage process, being sensitivity analysis 1 (SA1) and sensitivity analysis 2 (SA2). In the first stage (SA1), sensitivity of mutation and crossover was tested with a constant population size and constant number of generations. Those constant values for the population size and number of generations were cautiously selected based on preliminary testing, which was used initially to identify a reasonable range for the algorithm parameters (as previously reported in Table 1). The results obtained were visually compared with the well performing mutation and crossover operators then selected. In the second stage (SA2), these operators were applied for the sensitivity analysis of the population size and the number of generations. Additionally, all parameter setting combinations were analysed with both non-scaled and scaled objective functions.

Analysis name . | Sensitivity analysis of . | Setting of constant parameters . | Setting of variable parameters* . | No. of parameter setting combinations . | No. of analyses conducted . |
---|---|---|---|---|---|

SA1 | Crossover
| POP: 100 | MUT: 0.002; 0.005; 0.01; 0.1 | 40 | 40 (non-scaled**) |

Mutation
| GEN: 2,000 | CRS: 0.6; 0.7; 0.8; 0.9; 1.0 (all both S1P and SXP) | 40 (scaled**) | ||

SA2 | Population size
| MUT: 0.005 | POP: 50; 100; 150; 200; 300; 400 | 36 | 36 (non-scaled**) |

Number of generations
| CRS: 0.7 SXP; 1.0 S1P | GEN: 500; 1,000; 2,000 | 36 (scaled**) | ||

Total | 76 | 152 |

Analysis name . | Sensitivity analysis of . | Setting of constant parameters . | Setting of variable parameters* . | No. of parameter setting combinations . | No. of analyses conducted . |
---|---|---|---|---|---|

SA1 | Crossover
| POP: 100 | MUT: 0.002; 0.005; 0.01; 0.1 | 40 | 40 (non-scaled**) |

Mutation
| GEN: 2,000 | CRS: 0.6; 0.7; 0.8; 0.9; 1.0 (all both S1P and SXP) | 40 (scaled**) | ||

SA2 | Population size
| MUT: 0.005 | POP: 50; 100; 150; 200; 300; 400 | 36 | 36 (non-scaled**) |

Number of generations
| CRS: 0.7 SXP; 1.0 S1P | GEN: 500; 1,000; 2,000 | 36 (scaled**) | ||

Total | 76 | 152 |

POP = population size, GEN = number of generations, MUT = mutation, CRS = crossover.

*All possible combinations were analysed.

**Objective functions either non-scaled or scaled.

A single run was performed for each parameter setting combination from an initial population of random solutions. Therefore, the sensitivity of seed value (i.e. convergence of an algorithm from different initial populations of random solutions) was not investigated as part of this work and has been left for further research. A seed value of 1 was used for all analyses conducted.

Generally, to evaluate the sensitivity of a particular model parameter (i.e. model input), the following question needs to be answered: How sensitive is the solution (i.e. model output such as Pareto front) to the change of this particular parameter, when other parameters are kept constant? Analysing sensitivity of individual parameters, authors strive to answer this question. It is paramount to note that the sensitivity analysis was undertaken as performance evaluation. For example, if a particular parameter (adapting various values, when other parameters are kept constant) delivers ‘good’ solutions for some of its values and ‘poor’ solutions for others, a solution is very sensitive towards this parameter, and consequently the parameter is very sensitive. Vice versa, when a parameter keeps delivering ‘good’ solutions (while it adapts various values and other parameters are kept constant), it is not very sensitive.

Two measures were used for performance evaluation of parameter setting combinations; first, a visual comparison and second, a set of performance metrics. Due to the large amount of results, where often the Pareto fronts are non-parallel and intersecting, the visual comparison in isolation would become a very time consuming and difficult task. For this reason, the set of metrics supplemented the visual comparison to provide a faster, less subjective, and more quantitative means for performance evaluation.

Only Pareto fronts, in which all solutions are feasible (i.e. comply with constraints and thus are useful for operators), can be used for performance evaluation. The exclusion of Pareto fronts containing infeasible solutions is to ensure that a parameter setting combination, which delivered a Pareto front with infeasible solutions, cannot score better than another parameter setting combination, which produced only feasible solutions.

### Performance metrics

Performance metrics can measure algorithm effectiveness and efficiency (Van Veldhuizen 1999). In this work, effectiveness (or quality of solution) is the main aim as to evaluate how parameter settings affect the final solution. In other words, the quality of the Pareto front is the forefront interest, thus objective function values are being evaluated only. In order to keep consistent terminology with previous research, the following terms are adapted from Van Veldhuizen & Lamont (2000a):

A known Pareto front (PF

_{known}), which is the final computed Pareto front returned by the NSGA-II at termination, for the particular parameter setting combination.The true Pareto front (PF

_{true}), similarly described as a true Pareto set, ‘is not explicitly known for problems of any difficulty … is implicitly defined by the functions composing an MOP (multi-objective optimisation problem) – it is fixed and does not change’ (Van Veldhuizen & Lamont 2000a).

Similar to Baran *et al.* (2005), a very good approximation to the PF_{true} is obtained with the union of all computed PF_{knowns}. More specifically, PF_{true} used for this work is the union of the best solutions found in all analyses conducted for all parameter setting combinations, including both alternatives for objective function scaling.

No single metric can entirely capture total algorithm performance (Van Veldhuizen 1999; Deb *et al.* 2002). For this reason, a set of eight metrics is applied, which were adapted mainly from Van Veldhuizen (1999), Zitzler (1999) and Zitzler *et al.* (2000, 2003). Metrics were selected with the intent of being able to capture the overall quality of the PF_{known}. Some of these metrics compare similarity and proximity of a PF_{known} with the PF_{true} (i.e. how distant they are), whereas others evaluate the quality of a PF_{known} (i.e. the number of non-dominated solutions) or the quality of its individual solutions (i.e. spacing between these solutions). As PF_{known} is a result of a parameter setting combination, it is worth noting that by assessing the quality of the PF_{known}, the performance of the parameter setting combination can be evaluated.

Note that two major modifications were made to the original performance metrics of Van Veldhuizen (1999), Zitzler (1999) and Zitzler *et al.* (2000, 2003). Firstly, the original Van Veldhuizen and Zitzler metrics were designed for n-dimensional objective space. As two objectives are considered in this work, the metrics were converted to the (simpler) form relevant to two-dimensional objective space. Secondly, some metrics were transformed, where possible, into a normalised or relative value (i.e. per cent) which is suited better for the subsequent analysis.

#### Non-dominated number

_{known}. It is a modification of Van Veldhuizen's metric overall non-dominated vector generation (ONVG) with its original definition as (Van Veldhuizen 1999): where |

*PF*| is the cardinality of PF

_{known}_{known}, expressing the total number of solutions in PF

_{known}.

*POP*is the population size used to obtain the PF

_{known}.

It is expected that the NSGA-II with calibrated parameter settings is able to evolve the entire initial population of random solutions into the same number of non-dominated optimal solutions. It is, therefore, desirable that NN=100%, which means that the final number of non-dominated solutions equals to the population size.

#### Unique non-dominated number

_{known}. A unique solution represents a solution which has a different pump schedule from other solutions. The metric UN is defined as: where

*η*is a solution in PF

_{i}_{known},

*NN’*is the number of non-dominated solutions in PF

_{known}.

The reason for this metric is that NSGA-II may not always preserve the diversity of population, thus it may return some solutions in PF_{known} which are identical (i.e. having the same pump schedules). This means that some members of the initial population of random solutions evolved into the identical final solutions. Having identical solutions in the PF_{known} is not desirable, nonetheless, as diverse solutions are required to be available to a decision maker. In an ideal case, all solutions in PF_{known} are unique (i.e. UN=100%).

#### True number

_{known}, which are members of the PF

_{true}. This metric is a variation of the original Van Veldhuizen's metric error ratio (E) reporting the number of solutions in PF

_{known}, which are not members of the PF

_{true}(Van Veldhuizen 1999) as: A result E=0 (an ideal situation) indicates that all solutions in PF

_{know}are contained in the PF

_{true}, while E=1 indicates that there are none. For easier interpretation of how many solutions in PF

_{known}are part of the PF

_{true}, TN simply reports the percentage of such solutions as:

In an ideal situation, when all solutions in PF_{known} are members of PF_{true}, TN=100%. The higher the percentage of PF_{known} solutions are members of PF_{true}, the better the PF_{known} is.

#### Generational distance

_{known}to the PF

_{true}is, by measuring the distance between those two Pareto fronts. It is adapted from Van Veldhuizen (1999) as follows: where

*δ*is the Euclidean distance between the

_{i}*i*th solution in PF

_{known}and the nearest solution in PF

_{true}. For example, the Euclidean distance between the two points

*A=*(

*x*) and

_{A}, y_{A}*B=*(

*x*) in 2-dimensional space

_{B}, y_{B}*x-y*is calculated as

*δ*[(

_{AB}=*x*)

_{A}–x_{B}^{2}

*+*(

*y*)

_{A}–y_{B}^{2}]

^{1/2}.

The GD metric is kept as an absolute value consistent with the original Van Veldhuizen's formula (Van Veldhuizen 1999). It is obvious that the smaller the distance, the closer the PF_{know} is to the PF_{true}. A result GD=0 indicates that PF_{know}=PF_{true} in terms of location, any other value indicates that PF_{known} deviates from the PF_{true} (Van Veldhuizen 1999), so zero is a desired value.

ɛ -indicator

*ɛ*-indicator (IE) measures ‘the smallest distance that an approximation set [PF

_{known}] must be translated in order to completely dominate a reference set [PF

_{true}]’ (Kollat

*et al.*2008). This metric is adapted from Zitzler

*et al.*(2003) as: where is the value of the

*i*th objective in PF

_{known}, is the value of the

*i*th objective in PF

_{true}.

The IE metric adopts values equal or bigger than 1. A result IE=1 indicates that PF_{know}=PF_{true} in terms of location, the bigger IE value represents the larger distance from the PF_{true}, so 1 is a desired value.

#### S-metric

*et al.*2003). This metric measures the area or volume, respectively, covered by the PF

_{known}from the worst possible solution (the reference point). The SM is defined as follows: where

*a*is the area determined by the solutions in PF

_{i}_{known}and the reference point.

The SM is kept as an absolute value. A PF_{known} with the largest SM value is closest to the PF_{true}. Therefore, the larger value of SM, which expresses the closer proximity to the PF_{known}, is desirable.

#### Extent

_{known}across the objective space. This metric is a modification of Zitzler's metric (M

_{3}) calculated as (Zitzler 1999; Zitzler

*et al.*2000): where

*ɛ*is the Euclidean distance between the objective function values of two outer solutions in PF

_{outer}_{known}.

_{known}in relation to the PF

_{true}as follows: where is the Euclidean distance between the objective function values of two outer solutions in the PF

_{true}.

The extent of PF_{known} ‘should be maximised, i.e. for each objective a wide range of values should be covered by the non-dominated solutions’ (Zitzler *et al.* 2000), so a wide range of solutions is available to a decision maker. Equally, the greater value of EX expressing the larger the extent of PF_{known} is desirable. It is worth noting that the extent of a PF_{known}> 100% indicates that the PF_{known} has the larger extent than the PF_{true} (i.e. EX_{PFtrue}=100%).

#### Spacing

_{known}and is adapted from Van Veldhuizen (1999) as follows: where

*ɛ*is the Euclidean distance between the

_{i}*i*th solution and its closest neighbour in PF

_{known}, calculated in the objective space; is the mean of all

*ɛ*.

_{i}The SC metric is kept as an absolute value according to the original Van Veldhuizen's (1999) formula. To ensure that the results are not affected by identical solutions in the PF_{known}, only unique solutions were included into SC calculations. The more uniformly distributed solutions within the PF_{known}, the better the PF_{known} is. A result of SC = 0 indicates an equidistant spacing between solutions in the PF_{known}, which is an ideal situation.

An overview of the performance metrics, including their expected range and desired values, is shown in Table 3. Some of these metrics require knowledge about the PF_{true}, which is also indicated in Table 3.

Metric . | Abbreviation . | Expected range . | Desired value . | PF_{true} required?
. | Comment . |
---|---|---|---|---|---|

Non-dominated number | NN | 0–100% | 100% | No | Percentage of the population size |

Unique non-dominated number | UN | 0–100% | 100% | No | |

True number | TN | 0–100% | 100% | Yes | Percentage of the total number of non-dominated solutions in PF_{known} |

Generational distance | GD | ≥0 | 0 | Yes | Absolute value |

ɛ-indicator | IE | ≥1 | 1 | Yes (no)* | Absolute value |

S-metric | SM | ≥0 | >> 0 | No | Absolute value |

Extent | EX | ≥0% | >> 0% | Yes (no)** | Percentage of the PF_{true} extent |

Spacing | SC | ≥0 | 0 | No | Absolute value |

Metric . | Abbreviation . | Expected range . | Desired value . | PF_{true} required?
. | Comment . |
---|---|---|---|---|---|

Non-dominated number | NN | 0–100% | 100% | No | Percentage of the population size |

Unique non-dominated number | UN | 0–100% | 100% | No | |

True number | TN | 0–100% | 100% | Yes | Percentage of the total number of non-dominated solutions in PF_{known} |

Generational distance | GD | ≥0 | 0 | Yes | Absolute value |

ɛ-indicator | IE | ≥1 | 1 | Yes (no)* | Absolute value |

S-metric | SM | ≥0 | >> 0 | No | Absolute value |

Extent | EX | ≥0% | >> 0% | Yes (no)** | Percentage of the PF_{true} extent |

Spacing | SC | ≥0 | 0 | No | Absolute value |

*Individual PF_{knowns} could be compared without using PF_{true} (see Zitzler *et al.* 2003).

**If an absolute value was used (i.e. *EX = ε _{outer}*), PF

_{true}would not be required.

### Performance and sensitivity evaluation

There are two possible approaches to evaluate performance using the metrics. The first approach examines the results which were directly obtained from the metrics and compares them. This approach provides, on the one hand, very detailed information about performance of a parameter setting combination in each aspect of the quality of the Pareto front. On the other hand, the quantity of results (eight results for each parameter setting combination of 152 total combinations) would cause difficulties in analysing and presenting them. Hence, this approach was not used as it is more suitable for low quantities of results. It was used, for example, by Baran *et al.* (2005) to compare six multi-objective evolutionary algorithms.

The second approach, which is developed in the following paragraphs and applied for this work, combines the metrics together with the aim to obtain an overall performance for a parameter setting combination. This approach requires, besides setting the performance criteria for each of the metric with the corresponding performance scores, also the assignment of a weighting factor to each metric (Table 4). Based on the performance criteria, the value for each metric is replaced by means of linear interpolation with a corresponding performance score of 0.0–5.0 to provide a common basis for comparison. This step would not be required if all metrics were assessed as normalised or relative values. Weighting factors are then determined to provide a desired level of balance across all metrics according to the user's preference. These weighting factors (Table 4) combine the individual performance scores of metrics into a combined performance score of all metrics (Table 5). A disadvantage of this approach is in the subjectivity of weighting factors. However, the clear advantage is the significantly more succinct set of results obtained which are more easily interpreted and presented.

Metric . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM** . | EX (%) . | SC . | |
---|---|---|---|---|---|---|---|---|---|

Weighting factor . | 0.120 . | 0.050 . | 0.050 . | 0.250 . | 0.250 . | 0.250 . | 0.005 . | 0.025 . | |

Individual performance . | Performance score . | Metric performance criteria . | |||||||

Very good | 〈5.0–4.0) | 100–80 | 100–80 | 100–80 | 0–50 | 1–1.05 | 37.3*–34 | 313*–80 | 0–500 |

Good | 〈4.0–3.0) | 80–60 | 80–60 | 80–60 | 50–100 | 1.05–1.1 | 34–31 | 80–60 | 500–1,000 |

Average | 〈3.0–2.0) | 60–40 | 60–40 | 60–40 | 100–200 | 1.1–1.3 | 31–28 | 60–40 | 1,000–1,500 |

Poor | 〈2.0–1.0) | 40–20 | 40–20 | 40–20 | 200–500 | 1.3–1.5 | 28–25 | 40–20 | 1,500–2,000 |

Very poor | 〈1.0–0.0〉 | 20–0 | 20–0 | 20–0 | 500–12,816* | 1.5–3.34* | 25–21.3* | 20–0 | 2,000–48,838* |

Metric . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM** . | EX (%) . | SC . | |
---|---|---|---|---|---|---|---|---|---|

Weighting factor . | 0.120 . | 0.050 . | 0.050 . | 0.250 . | 0.250 . | 0.250 . | 0.005 . | 0.025 . | |

Individual performance . | Performance score . | Metric performance criteria . | |||||||

Very good | 〈5.0–4.0) | 100–80 | 100–80 | 100–80 | 0–50 | 1–1.05 | 37.3*–34 | 313*–80 | 0–500 |

Good | 〈4.0–3.0) | 80–60 | 80–60 | 80–60 | 50–100 | 1.05–1.1 | 34–31 | 80–60 | 500–1,000 |

Average | 〈3.0–2.0) | 60–40 | 60–40 | 60–40 | 100–200 | 1.1–1.3 | 31–28 | 60–40 | 1,000–1,500 |

Poor | 〈2.0–1.0) | 40–20 | 40–20 | 40–20 | 200–500 | 1.3–1.5 | 28–25 | 40–20 | 1,500–2,000 |

Very poor | 〈1.0–0.0〉 | 20–0 | 20–0 | 20–0 | 500–12,816* | 1.5–3.34* | 25–21.3* | 20–0 | 2,000–48,838* |

*Maximum or minimum values obtained.

**SM Metric performance criteria × 10^{6}.

Overall performance of a parameter setting combination . | Combined performance score . |
---|---|

Very good | 〈5.0–4.0) |

Good | 〈4.0–3.0) |

Average | 〈3.0–2.0) |

Poor | 〈2.0–1.0) |

Very poor | 〈1.0–0.0〉 |

Overall performance of a parameter setting combination . | Combined performance score . |
---|---|

Very good | 〈5.0–4.0) |

Good | 〈4.0–3.0) |

Average | 〈3.0–2.0) |

Poor | 〈2.0–1.0) |

Very poor | 〈1.0–0.0〉 |

The development of the metric performance criteria and weighting factors was undertaken by a trial-and-error approach. Initially, performance criteria for all metrics were distributed evenly with the same weighting factors assigned so all factors summed to 1.0. A process was then undertaken whereby the overall performance was assessed against a visual comparison where the proximity of the PF_{known} to the PF_{true} was considered of main importance. Basically, the visual comparison served as a means of ‘calibration’ of metric performance criteria and weighting factors, so the combined performance scores returned (see Table 5) were (more or less) in alignment with the location of the PF_{known} from the PF_{true}. For that reason, metrics GD, IE and SM were assigned a higher weighting factor than the other metrics reflecting their higher importance. Similarly, metric TN was assigned a low weighting factor because the bulk of the results (mainly SA1) for TN metric did not score well due to the lower population size applied. Fine tuning the performance criteria and weighting factors was an iterative task requiring careful considerations. Furthermore, these considerations will vary for different problems.

The following paragraphs illustrate a practical example of establishing the sensitivity of a particular parameter of interest using the performance metrics. Firstly, the metrics for the parameter setting combination are calculated and using linear interpolation, assigned a performance score from 0.0 to 5.0 (as per Table 4). For example, the results NN = 100%, UN = 94%, TN = 5%, GD = 117, IE = 1.06, SM = 37.3 × 10^{6}, EX = 99%, SC = 2,319 are replaced with performance scores as NN = 5.0, UN = 4.7, TN = 0.3, GD = 2.8, IE = 3.8, SM = 5.0, EX = 4.1, SC = 1.0. Secondly, the combined performance score, which can adopt values from 0.0 to 5.0, is established as the sum of individual metric performance scores added together after each of them was multiplied by a weighting factor. Accordingly, the combined performance score is obtained as 5.0 × 0.120 + 4.7 × 0.050 + 0.3 × 0.050 + 2.8 × 0.250 + 3.8 × 0.250 + 5.0 × 0.250 + 4.1 × 0.005 + 1.0 × 0.025 = 3.8, with the overall performance of the parameter setting combination ‘good’ (as per Table 5).

Finally, an additional step is undertaken to obtain the sensitivity of a particular parameter of interest. This involves grouping together parameter setting combinations, in which the parameter of interest is systematically changing (i.e. incrementally increasing) while the other parameters are held constant. These groups are further referred to as sensitivity groups (SEG). The sensitivity of the parameter of interest is then expressed in sensitivity points (SPs), which is the difference between the maximum and minimum combined performance scores within the SEG. For example, combined performance scores in the SEG ranging from 0.6 to 3.9 give 3.3 SPs (i.e. 3.9–0.6 = 3.3), indicating the high sensitivity of the parameter. Note that the maximum possible SPs = 5.0 (as per Table 5). Conversely, combined performance scores in the SEG ranging from 2.8 to 3.2 give 0.4 SPs, indicating low sensitivity of the parameter.

The methodology for performance and sensitivity analysis of algorithm parameters is summarised as a flowchart in Figure 3.

## RESULTS

All analyses were conducted using Intel Core i3 CPU M 380 processor with RAM 2.0 GB. Due to limited computer capacity for required analyses, about 30,000 function evaluations only were able to be run at once. Whereas the number of function evaluations required for one analysis ranged from 25,000 (i.e. population size of 50 for 500 generations) up to 800,000 (i.e. population size of 400 for 2,000 generations). Advantageously, the GANetXL's capability to conduct one analysis in a sequence of interrupted runs was used, where the previous run is resumed using the ‘resume’ function in the application's toolbar (CWS 2011).

The results for all 152 analyses conducted are displayed in Figure 4. Note that only feasible solutions are shown in Figure 4 as only feasible solutions were obtained from the optimisation model (6) subject to constraints (3)–(5). SA1 (solid grey squares in Figure 4), which tested the sensitivity of mutation and crossovers for a constant value of population size of 100 and the number of generations of 2000, includes 80 Pareto fronts (i.e. PF_{knowns}). SA2 (solid black triangles in Figure 4), which tested the sensitivity of the population size and number of generations for a constant value of mutation of 0.005 and crossover of 0.7 SXP and 1.0 S1P, includes 72 Pareto fronts (i.e. PF_{knowns}). It can be seen that SA1 (solid grey squares) consist of two isolated clouds of Pareto fronts. This is caused by a particular value of a mutation, which is discussed further in the Mutation section.

### True Pareto front

As previously described, PF_{true} is the union of the best solutions found in all analyses conducted. To be more exact, 152 Pareto fronts (i.e. PF_{knowns}) from both SA1 and SA2 were joined together (solid grey dots in Figure 5) and the optimal solutions identified (solid black dots as PF_{true} in Figure 5). Because some of these solutions were identical (i.e. had the same pump schedules), they were eliminated from the PF_{true}. Those identical solutions were mainly gained from the same combinations of NSGA-II parameter settings, but different objective function scaling. The resulting PF_{true}, which was used for subsequent calculations of some of the metrics, consists of a total of 333 non-dominated and unique solutions.

Some interesting observations were made regarding the PF_{true} (Figure 6, Table 6), which relate to the NSGA-II parameter setting combinations and population sizes creating the PF_{true}. The first observation was that the PF_{true} was reached by several NSGA-II parameter setting combinations, more specifically 13 combinations. Interestingly, five of those combinations (i.e. 38%) represented 91% of all solutions in PF_{true}. Conversely, eight combinations (i.e. 62%) represented only 9% of solutions in PF_{true}. So, the minority of the parameter setting combinations, which reached the PF_{true}, formed the vast majority of all PF_{true} solutions and vice versa. The second observation was that as the solutions within the PF_{true} progress from the top left to bottom right, they are derived from (more or less) dedicated sections to a particular and increasing population size. The last observation may indicate that population size performs inconsistently across a search space for this particular problem, which is discussed in the Discussion section.

NSGA-II parameter setting combination . | Solutions in PF_{true}. | ||||
---|---|---|---|---|---|

POP . | GEN . | MUT . | CRS . | Number . | % . |

50 | 2,000 | 0.005 | 1.0S1P | 1 | 0.3 |

100 | 2,000 | 0.005 | 1.0S1P | 10 | 3.0 |

100 | 2,000 | 0.01 | 0.7S1P | 15 | 4.5 |

100 | 2,000 | 0.005 | 1.0SXP | 2 | 0.6 |

150 | 2,000 | 0.005 | 0.7SXP | 47 | 14.1 |

150 | 1,000 | 0.005 | 0.7SXP | 4 | 1.2 |

200 | 2,000 | 0.005 | 0.7SXP | 2 | 0.6 |

300 | 2,000 | 0.005 | 0.7SXP | 133 | 39.9 |

300 | 2,000 | 0.005 | 1.0S1P | 30 | 9.0 |

300 | 1,000 | 0.005 | 0.7SXP | 3 | 0.9 |

300 | 1,000 | 0.005 | 1.0S1P | 2 | 0.6 |

400 | 2,000 | 0.005 | 1.0S1P | 79 | 23.7 |

400 | 1,000 | 0.005 | 1.0S1P | 5 | 1.5 |

Total | 333 | 100 |

NSGA-II parameter setting combination . | Solutions in PF_{true}. | ||||
---|---|---|---|---|---|

POP . | GEN . | MUT . | CRS . | Number . | % . |

50 | 2,000 | 0.005 | 1.0S1P | 1 | 0.3 |

100 | 2,000 | 0.005 | 1.0S1P | 10 | 3.0 |

100 | 2,000 | 0.01 | 0.7S1P | 15 | 4.5 |

100 | 2,000 | 0.005 | 1.0SXP | 2 | 0.6 |

150 | 2,000 | 0.005 | 0.7SXP | 47 | 14.1 |

150 | 1,000 | 0.005 | 0.7SXP | 4 | 1.2 |

200 | 2,000 | 0.005 | 0.7SXP | 2 | 0.6 |

300 | 2,000 | 0.005 | 0.7SXP | 133 | 39.9 |

300 | 2,000 | 0.005 | 1.0S1P | 30 | 9.0 |

300 | 1,000 | 0.005 | 0.7SXP | 3 | 0.9 |

300 | 1,000 | 0.005 | 1.0S1P | 2 | 0.6 |

400 | 2,000 | 0.005 | 1.0S1P | 79 | 23.7 |

400 | 1,000 | 0.005 | 1.0S1P | 5 | 1.5 |

Total | 333 | 100 |

### Objective function scaling

The results for the sensitivity of objective function scaling (Figure 7) include two sets of 76 Pareto fronts (i.e. PF_{knowns}) each gained from the same NSGA-II parameter setting combinations, with the difference that only one set had its objective functions scaled. Figure 7 indicates that there may be a considerable number of identical solutions in the two Pareto front sets, because black dots representing scaled objective functions appear to lie on top of grey crosses representing non-scaled objective functions. Comparing individual solutions from the two Pareto front sets confirmed that about 70% of them are indeed identical, resulting in the majority of Pareto fronts (64 of 76) from the two sets being also identical. Accordingly, 64 NSGA-II parameter setting combinations (84%) delivered the same Pareto fronts regardless of the objective functions being or not being scaled. Whereas only 12 NSGA-II parameter setting combinations (16%) delivered varying Pareto fronts for scaled and non-scaled objective functions.

Differences in those two sets of 12 Pareto fronts were each examined using metrics. Results from the SC metric showed that uniformity of spacing between solutions in the Pareto fronts with scaled objective functions improved in five instances and deteriorated in seven instances. Consequently, it was not confirmed that adequate scaling of the objective functions assures more uniformly distributed solutions, for the optimisation model and network used. Similarly, the improvement in quality of Pareto fronts for scaled objective functions regarding other performance metrics was non-conclusive, as some of those metrics gave better results and some worse, with the overall performance being the same comparing scaled and non-scaled objective functions. An exception was the EX metric, which for scaled objective functions improved in nine instances and deteriorated in three instances only, so scaled objective functions generally delivered Pareto fronts with larger extents.

Considering the bias of the NSGA-II towards the water quality objective when objective functions are inadequately scaled, the results for non-scaled and scaled objective functions were compared in terms of concentrations of solutions along the Pareto front and the location of the Pareto front. It can be seen from Figure 7 that the concentration of solutions for minimum pump energy cost values improved (for example ellipses A and B in Figure 7) for scaled objective functions. Moreover, scaled objective functions tended to find better solutions for minimum water quality values (for example ellipse E in Figure 7), with an example provided in Table 7. It was thus confirmed that non-scaled (i.e. inadequately scaled) objective functions caused the NSGA-II bias towards the water quality objective for the optimisation model and network used. However, the discontinuity of Pareto fronts is still obvious (for example ellipses C and D in Figure 7), with the reasons discussed in the Discussion section.

Analysis ID . | POP . | GEN . | MUT . | CRS . | Objective functions scaled . | Economic objective . | Water quality objective . |
---|---|---|---|---|---|---|---|

104 | 100 | 2,000 | 0.005 | 1.0 S1P | No | 148.7 | 98,149.0 |

204.8 | 1,815.9 | ||||||

140 | 100 | 2,000 | 0.005 | 1.0 S1P | Yes | 148.7 | 98,149.0 |

204.5 | 1,521.8* |

Analysis ID . | POP . | GEN . | MUT . | CRS . | Objective functions scaled . | Economic objective . | Water quality objective . |
---|---|---|---|---|---|---|---|

104 | 100 | 2,000 | 0.005 | 1.0 S1P | No | 148.7 | 98,149.0 |

204.8 | 1,815.9 | ||||||

140 | 100 | 2,000 | 0.005 | 1.0 S1P | Yes | 148.7 | 98,149.0 |

204.5 | 1,521.8* |

POP = population size, GEN = number of generations, MUT = mutation, CRS = crossover.

*Note that the water quality objective improved when scaled.

### NSGA-II parameters

Because the two sets of results for non-scaled and scaled objective functions were very similar, with scaled objective functions proved to deliver slightly better results, the subsequent sensitivity analysis of NSGA-II parameters was undertaken using results with scaled objective functions only.

The graphical results for exploring the sensitivity of mutation, crossover (both types S1P and SXP), population size and the number of generations are displayed in Figures 8–12. In those figures, each set of results representing a single parameter value (mutation value of 0.002 as white triangles in Figure 8, or crossover value of 0.6 as solid black squares in Figure 9, and so on) exist as a ‘cloud’ rather than a single front. The reason for this is that a single parameter value was always combined with other parameters, some as constants, but some as a range, as noted in the figure titles. For example, the data set for the aforementioned mutation value of 0.002 (white triangles in Figure 8) includes results for population size of 100, the number of generations of 2,000 and five crossover values from 0.6 to 1.0 both S1P and SXP, so this data set contains 10 Pareto fronts.

#### Mutation

The observation gleaned from Figure 8 is that the best and most competitive results were obtained from the mutation values of 0.005 and 0.01. The mutation value of 0.002 gave poorer results and the value of 0.1 provided Pareto fronts very distant from the optimal solutions, creating an isolated cloud. This indicates that the mutation value of 0.1 is too high for this particular problem, which caused the search to degenerate into a random process as previously described by Savic & Walters (1997). In contrast, the mutation value of 0.002 is too low for this particular problem, which caused the premature convergence of the algorithm to a local optimum as previously described by Srinivas & Patnaik (1994).

The sensitivity of mutation was investigated using metrics. Table 8 lists two sample SEGs from 10 available SEGs for the mutation. The performance of mutation varies from ‘very poor’ to ‘good’ for the both SEGs, with the corresponding 3.3 and 2.9 SPs, respectively. Because of the wide range of scores across all SEGs with the average of 2.9 SPs, the sensitivity of mutation is high for the optimisation model and network used. This means that the selection of a suitable mutation value to achieve good results requires careful attention.

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Mutation value . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

41 | 0.002 | POP: 100 GEN: 2,000 CRS: 0.6 S1P | 100 | 51 | 0 | 93 | 1.10 | 31.9 | 38 | 1,071 | 3.0 | Average |

51 | 0.005 | 100 | 86 | 0 | 125 | 1.10 | 32.3 | 63 | 1,914 | 2.9 | Average | |

61 | 0.01 | 100 | 91 | 0 | 61 | 1.06 | 35.4 | 45 | 1,335 | 3.9 | Good | |

71 | 0.1 | 18 | 18 | 0 | 4,113 | 1.74 | 21.3 | 149 | 14,416 | 0.6 | Very poor | |

48 | 0.002 | POP: 100 GEN: 2,000 CRS: 0.9 SXP | 100 | 57 | 0 | 83 | 1.17 | 28.7 | 45 | 1,759 | 2.7 | Average |

58 | 0.005 | 100 | 100 | 0 | 71 | 1.10 | 32.0 | 58 | 1,551 | 3.2 | Good | |

68 | 0.01 | 100 | 100 | 0 | 55 | 1.09 | 32.7 | 66 | 3,615 | 3.3 | Good | |

78 | 0.1 | 14 | 14 | 0 | 11,085 | 2.36 | 22.0 | 219 | 17,899 | 0.4 | Very poor |

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Mutation value . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

41 | 0.002 | POP: 100 GEN: 2,000 CRS: 0.6 S1P | 100 | 51 | 0 | 93 | 1.10 | 31.9 | 38 | 1,071 | 3.0 | Average |

51 | 0.005 | 100 | 86 | 0 | 125 | 1.10 | 32.3 | 63 | 1,914 | 2.9 | Average | |

61 | 0.01 | 100 | 91 | 0 | 61 | 1.06 | 35.4 | 45 | 1,335 | 3.9 | Good | |

71 | 0.1 | 18 | 18 | 0 | 4,113 | 1.74 | 21.3 | 149 | 14,416 | 0.6 | Very poor | |

48 | 0.002 | POP: 100 GEN: 2,000 CRS: 0.9 SXP | 100 | 57 | 0 | 83 | 1.17 | 28.7 | 45 | 1,759 | 2.7 | Average |

58 | 0.005 | 100 | 100 | 0 | 71 | 1.10 | 32.0 | 58 | 1,551 | 3.2 | Good | |

68 | 0.01 | 100 | 100 | 0 | 55 | 1.09 | 32.7 | 66 | 3,615 | 3.3 | Good | |

78 | 0.1 | 14 | 14 | 0 | 11,085 | 2.36 | 22.0 | 219 | 17,899 | 0.4 | Very poor |

POP = population size, GEN = number of generations, CRS = crossover.

#### Crossover

It can be observed from Figures 9 and 10 that particular crossover values are very widely spread and, unlike mutation, are not formed into designated performance areas. This may not make the identification of the best performing crossover value as straightforward as for the mutation. Indeed, the metrics with the following sensitivity analysis confirmed the low sensitivity to different crossover values, meaning that the change of crossover setting does not affect the final solution significantly.

Table 9 presents two sample SEGs from eight available SEGs for the crossover. The performance of crossover varies from ‘average’ to ‘good’ for the first SEG and is ‘very poor’ across the second SEG, with the corresponding 0.9 SPs and 0.4 SPs, respectively. Because the range of scores across all SEGs is narrow with the average of 0.7 SPs, the sensitivity of crossover is low for the optimisation model and network used. The selection of a suitable crossover value for this particular problem is still important to achieve good results, yet it does not seem as crucial as the identification of a suitable mutation value.

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Crossover value . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

51 | 0.6s1p | POP: 100 GEN: 2,000 MUT: 0.005 | 100 | 86 | 0 | 125 | 1.10 | 32.3 | 63 | 1,914 | 2.9 | Average |

53 | 0.7s1p | 100 | 96 | 0 | 59 | 1.13 | 33.0 | 52 | 1,539 | 3.2 | Good | |

55 | 0.8s1p | 100 | 97 | 0 | 58 | 1.08 | 33.8 | 40 | 1,245 | 3.3 | Good | |

57 | 0.9s1p | 100 | 70 | 0 | 45 | 1.14 | 29.1 | 46 | 2,344 | 2.9 | Average | |

59 | 1.0s1p | 100 | 94 | 5 | 117 | 1.06 | 37.3 | 99 | 2,319 | 3.8 | Good | |

72 | 0.6sXp | POP: 100 GEN: 2,000 MUT: 0.1 | 16 | 16 | 0 | 5,655 | 1.84 | 22.9 | 166 | 14,186 | 0.6 | Very poor |

74 | 0.7sXp | 18 | 18 | 0 | 12,816 | 3.34 | 24.0 | 313 | 48,838 | 0.4 | Very poor | |

76 | 0.8sXp | 17 | 17 | 0 | 3,775 | 1.42 | 22.8 | 123 | 7,396 | 0.8 | Very poor | |

78 | 0.9sXp | 14 | 14 | 0 | 11,085 | 2.36 | 22.0 | 219 | 17,899 | 0.4 | Very poor | |

80 | 1.0sXp | 16 | 16 | 0 | 7,259 | 1.83 | 23.6 | 167 | 13,545 | 0.6 | Very poor |

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Crossover value . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

51 | 0.6s1p | POP: 100 GEN: 2,000 MUT: 0.005 | 100 | 86 | 0 | 125 | 1.10 | 32.3 | 63 | 1,914 | 2.9 | Average |

53 | 0.7s1p | 100 | 96 | 0 | 59 | 1.13 | 33.0 | 52 | 1,539 | 3.2 | Good | |

55 | 0.8s1p | 100 | 97 | 0 | 58 | 1.08 | 33.8 | 40 | 1,245 | 3.3 | Good | |

57 | 0.9s1p | 100 | 70 | 0 | 45 | 1.14 | 29.1 | 46 | 2,344 | 2.9 | Average | |

59 | 1.0s1p | 100 | 94 | 5 | 117 | 1.06 | 37.3 | 99 | 2,319 | 3.8 | Good | |

72 | 0.6sXp | POP: 100 GEN: 2,000 MUT: 0.1 | 16 | 16 | 0 | 5,655 | 1.84 | 22.9 | 166 | 14,186 | 0.6 | Very poor |

74 | 0.7sXp | 18 | 18 | 0 | 12,816 | 3.34 | 24.0 | 313 | 48,838 | 0.4 | Very poor | |

76 | 0.8sXp | 17 | 17 | 0 | 3,775 | 1.42 | 22.8 | 123 | 7,396 | 0.8 | Very poor | |

78 | 0.9sXp | 14 | 14 | 0 | 11,085 | 2.36 | 22.0 | 219 | 17,899 | 0.4 | Very poor | |

80 | 1.0sXp | 16 | 16 | 0 | 7,259 | 1.83 | 23.6 | 167 | 13,545 | 0.6 | Very poor |

POP = population size, GEN = number of generations, MUT = mutation.

#### Population size

The performance of different population sizes (Figures 6 and 11) was discussed in relation to the PF_{true} in previous sections, so only sensitivity is examined here. Table 10 lists two sample SEGs from six available SEGs for the population size. The performance of population size varies from ‘average’ to ‘very good’ for the first SEG and is ‘good’ across the second SEG, with the corresponding 1.4 SPs and 0.7 SPs, respectively. Because of the intermediate range of scores across all SEGs with the average of 1.2 SPs, the sensitivity of population size is moderate for the optimisation model and network used. The selection of a suitable population size to achieve good results needs careful attention, which is emphasised by the finding that a particular population size appears to perform better at some locations within the search space and worse in others as demonstrated by the PF_{true} (Figure 6). This inconsistent performance is discussed in the Discussion section.

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Population size . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

118 | 50 | GEN: 1,000 MUT: 0.005 CRS: 0.7 SXP | 100 | 100 | 0 | 202 | 1.13 | 32.1 | 83 | 4,119 | 2.8 | Average |

121 | 100 | 100 | 100 | 0 | 23 | 1.08 | 30.6 | 22 | 927 | 3.2 | Good | |

124 | 150 | 100 | 86 | 3 | 46 | 1.05 | 36.5 | 64 | 1,239 | 4.1 | Very good | |

127 | 200 | 100 | 51 | 0 | 42 | 1.12 | 35.1 | 54 | 867 | 3.7 | Good | |

130 | 300 | 100 | 48 | 2 | 36 | 1.12 | 35.8 | 71 | 1,219 | 3.7 | Good | |

133 | 400 | 100 | 43 | 0 | 102 | 1.15 | 33.4 | 115 | 2,007 | 2.7 | Average | |

137 | 50 | GEN: 2,000 MUT: 0.005 CRS: 1.0 S1P | 100 | 98 | 2 | 67 | 1.10 | 31.8 | 55 | 2,474 | 3.2 | Good |

140 | 100 | 100 | 94 | 5 | 117 | 1.06 | 37.3 | 99 | 2,319 | 3.8 | Good | |

143 | 150 | 100 | 99 | 0 | 24 | 1.13 | 33.7 | 38 | 1,139 | 3.3 | Good | |

146 | 200 | 100 | 100 | 0 | 29 | 1.11 | 31.3 | 48 | 988 | 3.5 | Good | |

149 | 300 | 100 | 34 | 7 | 24 | 1.07 | 34.6 | 58 | 822 | 3.9 | Good | |

152 | 400 | 100 | 83 | 7 | 22 | 1.08 | 32.9 | 53 | 1,017 | 3.5 | Good |

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Population size . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

118 | 50 | GEN: 1,000 MUT: 0.005 CRS: 0.7 SXP | 100 | 100 | 0 | 202 | 1.13 | 32.1 | 83 | 4,119 | 2.8 | Average |

121 | 100 | 100 | 100 | 0 | 23 | 1.08 | 30.6 | 22 | 927 | 3.2 | Good | |

124 | 150 | 100 | 86 | 3 | 46 | 1.05 | 36.5 | 64 | 1,239 | 4.1 | Very good | |

127 | 200 | 100 | 51 | 0 | 42 | 1.12 | 35.1 | 54 | 867 | 3.7 | Good | |

130 | 300 | 100 | 48 | 2 | 36 | 1.12 | 35.8 | 71 | 1,219 | 3.7 | Good | |

133 | 400 | 100 | 43 | 0 | 102 | 1.15 | 33.4 | 115 | 2,007 | 2.7 | Average | |

137 | 50 | GEN: 2,000 MUT: 0.005 CRS: 1.0 S1P | 100 | 98 | 2 | 67 | 1.10 | 31.8 | 55 | 2,474 | 3.2 | Good |

140 | 100 | 100 | 94 | 5 | 117 | 1.06 | 37.3 | 99 | 2,319 | 3.8 | Good | |

143 | 150 | 100 | 99 | 0 | 24 | 1.13 | 33.7 | 38 | 1,139 | 3.3 | Good | |

146 | 200 | 100 | 100 | 0 | 29 | 1.11 | 31.3 | 48 | 988 | 3.5 | Good | |

149 | 300 | 100 | 34 | 7 | 24 | 1.07 | 34.6 | 58 | 822 | 3.9 | Good | |

152 | 400 | 100 | 83 | 7 | 22 | 1.08 | 32.9 | 53 | 1,017 | 3.5 | Good |

GEN = number of generations, MUT = mutation, CRS = crossover.

#### Number of generations

The first observation gleaned from Figure 12 is that generally, the results show steady improvement with the increasing number of generations. Hence, the best results were obtained from the number of generations of 2,000, which was confirmed by the PF_{true}, where 319 (i.e. 96%) of total 333 solutions in the PF_{true} were gained from this number of generations.

The sensitivity of the number of generations was assessed using metrics. Table 11 lists two sample SEGs from 12 available SEGs for the number of generations. The performance of the number of generations varies from ‘average’ to ‘good’ for the first SEG and is ‘good’ across the second SEG, with the corresponding 0.4 SPs and 0.8 SPs, respectively. Because the range of scores across all SEGs is narrow with the average of 0.7 SPs, the sensitivity of number of generations is low for the optimisation model and network used. In spite of the steady improvements in results with increased number of generations, there are additional associated computational costs which may be considered to negate the perceived benefits.

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Number of generations . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

120 | 500 | POP: 100 MUT: 0.005 CRS: 0.7 SXP | 100 | 71 | 0 | 72 | 1.14 | 29.5 | 52 | 2,764 | 2.8 | Average |

121 | 1,000 | 100 | 100 | 0 | 23 | 1.08 | 30.6 | 22 | 927 | 3.2 | Good | |

122 | 2,000 | 100 | 100 | 0 | 28 | 1.07 | 30.8 | 22 | 526 | 3.2 | Good | |

147 | 500 | POP: 300 MUT: 0.005 CRS: 1.0 S1P | 100 | 33 | 0 | 43 | 1.14 | 32.5 | 40 | 675 | 3.1 | Good |

148 | 1,000 | 100 | 28 | 0 | 51 | 1.08 | 34.3 | 59 | 1,001 | 3.6 | Good | |

149 | 2,000 | 100 | 34 | 7 | 24 | 1.07 | 34.6 | 58 | 822 | 3.9 | Good |

. | . | . | Metrics . | . | . | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|

ID . | Number of generations . | Constant parameters . | NN (%) . | UN (%) . | TN (%) . | GD . | IE . | SM (×10^{6})
. | EX (%) . | SC . | Combined performance score . | Overall performance . |

120 | 500 | POP: 100 MUT: 0.005 CRS: 0.7 SXP | 100 | 71 | 0 | 72 | 1.14 | 29.5 | 52 | 2,764 | 2.8 | Average |

121 | 1,000 | 100 | 100 | 0 | 23 | 1.08 | 30.6 | 22 | 927 | 3.2 | Good | |

122 | 2,000 | 100 | 100 | 0 | 28 | 1.07 | 30.8 | 22 | 526 | 3.2 | Good | |

147 | 500 | POP: 300 MUT: 0.005 CRS: 1.0 S1P | 100 | 33 | 0 | 43 | 1.14 | 32.5 | 40 | 675 | 3.1 | Good |

148 | 1,000 | 100 | 28 | 0 | 51 | 1.08 | 34.3 | 59 | 1,001 | 3.6 | Good | |

149 | 2,000 | 100 | 34 | 7 | 24 | 1.07 | 34.6 | 58 | 822 | 3.9 | Good |

POP = population size, MUT = mutation, CRS = crossover.

## DISCUSSION

### Performance metrics and their application

The UN metric (9) reports the percentage of unique non-dominated solutions in a PF_{known} (i.e. solutions with unique pump schedules). There were some solutions found which had different pump schedules, yet objective function values for both objectives were nearly the same, for example, only differing in the third decimal place for one objective. Note that solutions with the same objective function values, but different decision variables (i.e. pump schedules) are called alternate solutions (Bhattacharjya 2007). In this work, the formerly described solutions were considered ‘unique’ for the following reason. If a solution has desirable objective function values but is impractical, the decision maker would probably be interested in a different solution with similar objective function values.

There is a limitation to the GD metric (12), which measures the distance between a PF_{known} and the PF_{true}. This metric may give misleading results if the length (i.e. extent) of the PF_{known} is significantly greater than the extent of the PF_{true} as demonstrated in Figure 13. Despite the proximity of PF_{knownA} and PF_{knownB} to the PF_{true} being the same, the GD metric returns larger value for the PF_{knownA} due to inclusion of the points A1–A4 in the equation, calculating their distance from the outer PF_{true} point T1. This limitation was recognised but was considered adequate for the purpose of the sensitivity analysis. This limitation is overcome by using the *ɛ*-indicator metric.

In addition, both the GD metric and *ɛ*-indicator metric require knowledge about the PF_{true}, which is not normally available. The metric with a similar function, which measures the closeness of a PF_{known} to the PF_{true}, yet does not require the PF_{true} information, is S-metric for 2-dimensional objective space (Zitzler & Thiele 1998) or hypervolume for n-dimensional objective space (Fleischer 2003; Zitzler *et al.* 2003). The disadvantage of this metric is, however, that the choice of the reference point can influence the evaluation of the PF_{knowns} (Knowles & Corne 2002).

The methodology for performance and sensitivity analysis using metrics may be applied for the calibration of multi-objective algorithm parameters for any WDS optimisation problem. It needs to be noted, nevertheless, that the metric performance criteria, performance scores and weighting factors (see Table 4) were specifically developed for the purpose of this work and, that these parameters need to be developed and fine-tuned to suit the particular problem. Furthermore, the selection of the metrics and their performance evaluation also depends on the purpose for their application. In the case of calibration, the metrics GD, IE, SM and TN would likely adopt the higher weighting factors than the other metrics. For applications comprising more than two objectives, metrics can easily be adjusted to the required n-dimensional format.

The disadvantage of the methodology used is that a large computational effort is required. This effort may be partially reduced by using only metrics, which do not require knowledge about the PF_{true} (refer to Table 3). Alternatively, algorithms with adaptive parameter settings, such as Borg multi-objective evolutionary algorithm (Hadka & Reed 2013) or algorithms with few parameters, such as Pareto archived dynamically dimensioned search (PA-DDS) with only one parameter (Asadzadeh & Tolson 2011) could be considered. It is expected that these algorithms may well require less computational effort to calibrate their parameters.

### Objective function scaling

In spite of the adequate scaling of objective functions, the Pareto fronts obtained were discontinuous, which is possibly due to the following:

Discrete decision variables.

Model simplifications, for example, water quality analysis (the length of time step, i.e. 1 hour rather than 1–10 minutes, the length of simulation period, i.e. 1 day rather than 1 week or more).

Only a single run was performed from one initial population of random solutions.

### NSGA-II parameters

#### Parameter sensitivity

Performance metrics enabled to evaluate the sensitivity of the NSGA-II parameters on a quantitative and more objective basis than the use of visual comparison only. In particular, the sensitivity of NSGA-II parameters was expressed as SPs of maximum possible SPs, which assisted in understanding of sensitivity hierarchy (Table 12). The most sensitive parameter was identified mutation and the least sensitive crossover, with the limitation to the optimisation model and network used. Population size was more sensitive than the number of generations for this problem. Note that the more sensitive parameter is, its change has greater influence on results.

. | Sensitivity points (SPs)* . | . | . | |||
---|---|---|---|---|---|---|

Parameter . | Average . | Median . | Minimum . | Maximum . | Parameter sensitivity . | Number of sensitivity groups (SEGs) . |

Mutation | 2.9 | 2.9 | 2.7 | 3.3 | High | 10 |

Population size | 1.2 | 1.3 | 0.6 | 1.8 | Moderate | 6 |

Number of generations | 0.7 | 0.8 | 0.0 | 2.1 | Low | 12 |

Crossover | 0.7 | 0.7 | 0.2 | 1.4 | Low | 8 |

. | Sensitivity points (SPs)* . | . | . | |||
---|---|---|---|---|---|---|

Parameter . | Average . | Median . | Minimum . | Maximum . | Parameter sensitivity . | Number of sensitivity groups (SEGs) . |

Mutation | 2.9 | 2.9 | 2.7 | 3.3 | High | 10 |

Population size | 1.2 | 1.3 | 0.6 | 1.8 | Moderate | 6 |

Number of generations | 0.7 | 0.8 | 0.0 | 2.1 | Low | 12 |

Crossover | 0.7 | 0.7 | 0.2 | 1.4 | Low | 8 |

*Difference between the maximum and minimum combined performance scores within SEGs.

A comment needs to be made though that the identification of reasonable range of settings for each parameter was important in this exercise. Authors acknowledge that the value for the mutation of 0.1 may be a little too high and if this value was not used, the sensitivity of mutation would not be evaluated as high. However, quite a wide range of settings for each parameter was identified purposely so the impact of those settings on the final results could be clearly demonstrated.

#### Parameter interdependency

The performance of a mutation in relation to crossover and vice versa is now examined by exploring charts detailing concrete mutation and crossover values, and crossover types. Figure 14 indicates that an increase in mutation from 0.005 to 0.01 using the crossover of 0.6 improved the solution (arrow 1a in Figure 14); conversely, the solution deteriorated for the same increase in mutation, but crossover of 1.0 (arrow 1b in Figure 14). Vice versa, when crossover increased from 0.6 to 1.0 using mutation of 0.005, the solution improved (arrow 2a in Figure 14); conversely, the solution deteriorated for the same increase in crossover, but mutation of 0.01 (arrow 2b in Figure 14). Figure 15 displays Pareto fronts for varying mutation and crossover values, and crossover types. It demonstrates that the change of crossover type from S1P to SXP for crossover value of 0.6 and mutation value of 0.005 resulted in better results (arrow 3a in Figure 15); and vice versa, change of crossover type from S1P to SXP for crossover value of 1.0 and mutation value of 0.01 resulted in worse results (arrow 3b in Figure 15). The likely conclusion drawn from those observations is that the performance of a particular mutation depends on the crossover value and crossover type, and the performance of a particular crossover value and its type is probably dependent upon the mutation value. In other words, it appears that mutation and crossover are somehow interdependent. If this is the case, they ought to be evaluated as an interdependent pair rather than in isolation. More research is needed in this area. Plotting the values of computed metrics for different combinations of crossover and mutation operators on a surface or contour chart is suggested as one way of visualizing the combined effects of these two parameters.

#### Performance inconsistency

Due to Pareto fronts for different parameter setting combinations being non parallel often intersecting (Figures 14–16), it indicates that some parameter setting combinations may perform better at some locations of the search space and worse in others. At Figures 14 and 15, for example, mutation value of 0.005 and crossover value of 0.6 S1P (solid grey dots) perform well towards minimum water quality values, but quite irregularly diverge from minimum pump energy costs values. Another example is in Figure 16, where all population sizes were able to capture (more or less) minimum water quality values, but their level of performance (or accuracy) over the large search space differs greatly. Hence, (certain) parameter setting combinations seem to perform with various degrees of inconsistency across the search space. This inconsistent performance was also demonstrated by the PF_{true} (Figure 6), which was not reached by one but several parameter setting combinations. This is thought to reflect the complexity of the search space, and requires further research.

Unlike single-objective approaches, where the calibration process results in the ‘best’ algorithm parameter values to find the optimal solution as demonstrated in Younis & Dandy (2012), the identification of those ‘best’ values in multi-objective approach does not seem as straightforward. On the contrary, it appears that the multi-objective approach requires a range of parameter values (i.e. a range of parameter setting combinations) to reach or identify all optimal solutions within the (true) Pareto front. This may be caused by the fact that the single-objective approach seeks one particular solution, whereas the multi-objective approach searches for an entire front of optimal solutions.

### Limitations

There are three limitations to this work. The first limitation is that, due to the large number of parameter setting combinations being evaluated, only a single run of the NSGA-II was performed for each parameter setting combination. This approach is contrary to the common application of genetic algorithms (GAs) where multiple runs are used because of the stochastic nature of GAs. The reader, therefore, needs to keep in mind that the analyses of the generated results and conclusions are limited to a single run only (i.e. are not based on statistical assessment of multiple runs). The second limitation is that the number of parameter setting combinations to be evaluated needs to be limited (i.e. based on literature values and/or preliminary testing), otherwise time required to run analyses may become unacceptable. The third limitation is that one example network was used only. These limitations need to be addressed in future research.

## CONCLUSION

In this work, performance metrics were applied to comprehensively analyse the sensitivity of NSGA-II parameters and their settings and objective function scaling for multi-objective optimisation problem in WDSs. Algorithm parameters considered included population size, number of generations, crossover and mutation. A two-objective optimisation model involved the competing operational objectives of pump energy costs and water quality, and was applied to a test network adopted from the literature representing a regional non-drinking WDS.

Performance metrics enabled the evaluation of the sensitivity of the NSGA-II parameters on a quantitative and more objective basis than the use of visual comparison only. Moreover, they assisted in determining the sensitivity hierarchy of NSGA-II parameters for the particular optimisation model and network used. Performance metrics were thus demonstrated suitable for sensitivity analysis of NSGA-II parameters for multi-objective optimisation problems in WDSs. Additionally, the methodology for performance and sensitivity analysis may be used by researchers and practitioners, after converting metrics to the adequate n-dimensional objective space, for the calibration of multi-objective algorithm parameters for any WDS optimisation problem. In order to repeat a process, it is summarised in the form of a flowchart in Figure 3.

Adequate scaling of the objective functions to cover similar ranges of numerical values was confirmed important for the optimisation model and network used, because inadequately scaled objective functions caused the NSGA-II to bias towards the water quality objective. However, adequately scaled objective functions did not assure continuity of Pareto fronts. In fact, the discontinuity of Pareto fronts is likely attributed to other factors, for example discrete nature of the problem.

NSGA-II parameters were found sensitive for the optimisation model and the network used. Performance of the algorithm and the quality of the solution was thus dependent on parameter settings. Furthermore, some parameters (i.e. mutation and crossover) appeared interdependent and some parameter setting combinations performing with various degrees of inconsistency across the search space. It appears that unlike a single-objective approach, a multi-objective approach requires a range of parameter values (i.e. a range of parameter setting combinations) to cover the entire search space and to reach or identify all optimal solutions. However, additional future research is required in this area to confirm the results and conclusions of this work by the inclusion of different seed values into performance and sensitivity analysis of algorithm parameters, and using different applications in multi-objective optimisation problems of WDSs.

## ACKNOWLEDGEMENT

This research was supported by the Australian Research Council as Project LP0990908.