Abstract
The performance of model-based leak detection and localization techniques heavily depends on the configuration of a limited number of sensors. This paper presents a sensor placement optimization strategy that guarantees sufficient diagnosability while satisfying the budget constraint. Based on the theory of compressed sensing, the leak localization problem could be transformed into acquiring the sparse leak-induced demands from the available measurements, and the average mutual coherence is devised as a diagnosability criterion for evaluating whether the measurements contain enough information for identifying the potential leaks. The optimal sensor placement problem is then reformulated as a {0, 1} quadratic knapsack problem, seeking an optimal sensor placement scheme by minimizing the average mutual coherence to maximize the degree of diagnosability. To effectively handle the complicated real-life water distribution networks, a validated binary version of artificial bee colony algorithm enhanced by genetic operators, including crossover and swap, is introduced to solve the binary knapsack problem. The proposed strategy is illustrated and validated through a real-life water distribution network with synthetically generated field data.
INTRODUCTION
Water loss management is one of the most significant issues facing urban water utilities around the world. According to the World Bank study (Kingdom et al. 2006), a total amount of 45 million cubic meters of water are lost daily in developing countries, enough to serve nearly 200 million people, and causing a conservative estimate of US$4.7 billion economic losses per year. Among those losses, huge volumes of water are lost through leaks, not being effectively invoiced to customers, as a result of damage or defects in pipelines, lack of maintenance, or increase in pressure. Therefore, detection and localization of water leakages within water distribution networks (WDNs) is extremely urgent for the sustainable development of urban cities.
The past few years have witnessed much progress towards leak detection, localization and pinpointing. In this field, Hao et al. (2012) provides a comprehensive review of the state-of-the-art technologies for condition assessment of underground utilities, especially water supply pipelines. Unfortunately, although effective, these techniques, including widely used acoustic and vibration approaches (Gao et al. 2005), are time-consuming and resource-intensive when applied to large-scale WDNs. An alternative is to deploy the WDNs with flow and pressure monitoring devices (Puust et al. 2010). Generally, pressure instruments are preferred in practice due to the convenience for installation. Data-driven as well as model-based approaches are adopted, both with their respective advantages. Data-driven approaches barely rely on prior knowledge about the monitored WDNs, since they focus on extracting features of historical data (Bakker et al. 2014; Laucelli et al. 2016). Note that applying data-driven approaches may lead to misjudgements when inexperienced events are dealt with. The application of hydraulic transient models to leak localization is another hotspot in the research field. Colombo et al. (2009) offer a selective literature review of transient model-based methods. So far, even with long operational range and high positioning accuracy, the transient-based methods are not yet competent for supplanting existing techniques, due to their strong reliance on a number of pressure measuring instruments with high sampling frequency. On the other hand, leak detection and localization techniques on the basis of stationary models has been researched since the first seminal paper of Pudar & Liggett (1992). According to the theory of model-based fault detection and isolation (FDI), residuals online between pressure measurements and estimations using a hydraulic simulator are generated and evaluated. To detect leaks, residuals are checked against thresholds. Further, when some of the residuals violate the thresholds, the residuals are matched against the leak sensitivity matrix to locate potential leaks (Casillas et al. 2014; Rosich et al. 2015). The stationary model-based methods or mixed model-based/data-driven approaches overall offer more effective and less costly system-wide diagnosis and monitoring schemes (Meseguer et al. 2014; Zhang et al. 2016).
It is obvious that the location of pressure instrumentations is crucial to provide good quality information for optimizing the performance of different management tasks. Many studies have been conducted for optimal sensor placement, driven by the need for model calibration, network segmentation and leak localization respectively. To put it simply, the optimal sensor placement can be regarded as a multi-objective optimization problem that selectively balances the need for improving the performance with the need to justify the expense of data collection (including metering, installation and maintenance costs). For model calibration, the overall performance is evaluated by the estimation or prediction accuracy, quantified by three most often adopted indexes, the so-called A, D and V-optimality criteria (Savic et al. 2009), or combined (Kang & Lansey 2010). The A- and D-optimality are concerned with the estimated parameter uncertainty, aimed at either minimizing the trace or maximizing the determinant of the parameter covariance matrix, while the V-optimality is concerned with the prediction uncertainty, aimed at minimizing the trace of the prediction covariance matrix. For network segmentation, the sampling-oriented modularity (Simone et al. 2016) is employed as a segmentation metric to gain better ‘pressure DMAs’ configuration. For leak localization, various quantitative indexes are introduced for evaluating the sensor placement schemes on locating possible leaks. Christodoulou et al. (2013) define a total entropy describing the ratio of sensing length over pipe length, and optimizes the sensor placement by maximizing the total entropy in the network; Casillas et al. (2015) adopt a leak signature space (LSS) method, and employ the criterion of minimizing the number of overlapping signature domain using the integer optimization method; Sarrate et al. (2014) propose an isolability index as the number of isolable node pairs under a single operating point scenario, and search for optimal sensor placement by a branch and bound method based on structural analysis; Blesa et al. (2016) reformulate the sensor placement problem as a bi-objectives optimization for maximizing both mean and worst leak locatability index, and take into account the dependency of leak localization on sequential operating points; Steffelbauer & Fuchs-Hanusch (2016) comprehensively incorporate uncertainties from different types and sources in the optimal sensor placement problem. To sum up, the key in sensor placement optimization for leak localization is introducing a quantitative index to evaluate the quality of every possible placement scheme, for which a decision-maker would choose a preferred one.
This paper formulates the leak localization problem as a FDI problem with underdetermined linear equations, and adopts a diagnosability criterion to guide the sensor placement optimization so that leaks can be located. By maximizing the diagnosability using an enhanced binary artificial bee colony (ABC) algorithm, a quasi-optimal sensor placement can be guaranteed. The proposed sensor placement optimization approach is illustrated through a real-life network with synthetically generated field data.
DIAGNOSABILITY CRITERION AND OPTIMAL SENSOR PLACEMENT
In this section, a compressed sensing based diagnosability criterion is adopted to guide the sensor placement optimization and enhance the overall performance for leak detection and localization. The following subsection proposes a summary of the model-based leak detection and localization procedure. Particularly, an analytical expression for the leak sensitivity matrix is provided. Next, diagnosability analysis is presented for a leak localization procedure with the theory of compressed sensing. Average mutual coherence, a typical diagnosability criterion, is adopted for evaluating the quality of sensor placement. This is followed by a subsection describing the implementation of an enhanced binary ABC algorithm to support the decision making process, aimed at maximizing the degree of diagnosability (minimizing the average mutual coherence in this paper) and thus optimizing the sensor placement.
Leak detection and localization in WDNs
Model-based leak detection and localization techniques are adopted in this paper, aimed at providing a more effective and less costly system-wide diagnosis and monitoring scheme for WDNs. According to the theory of FDI, a set of online residuals between available pressure measurements and corresponding estimations using hydraulic simulators are generated and evaluated. In the absence of leaks, all of the generated residuals remain below given thresholds. Otherwise, when some of the residuals violate the thresholds, the residuals are matched against the leak sensitivity matrix, searching for the most favorable combination of leak candidates that leads to the consistency of the observed and expected residuals.
Compressed sensing based diagnosability analysis
In compressed sensing, diagnosability is a reasonable measure for evaluating whether the measurements contain enough information for exact recovery of every possible sparse vector l, regardless of the practicality of the reconstruction scheme. So far, the diagnosability criterion has been under extensive research and is manifested as some requirements on ΛS (Bastani et al. 2016). As the leak sensitivity matrix S is specific according to the network topology and operating point scenario, a better measurement matrix Λ is thus desired, corresponding to a better sensor placement scheme that leads to better compressed-sensing reconstruction performance.
Optimal sensor placement using enhanced binary ABC algorithm
Ozturk et al. (2015) demonstrate that the enhanced binary ABC algorithm outperforms other analogous algorithms in solving the binary knapsack problem. The ABC algorithm is a population-based metaheuristic technique, motivated by the intelligent foraging behavior of a honey bee swarm, for effectively solving non-convex and non-smooth optimization problem. The standard ABC algorithm can effectively optimize continuous problems, and for taking advantage of the strong find-best ability of the ABC algorithm, some modifications on ABC are required in order to suit the binary optimization for the knapsack problem. Eventually, it should be successfully applied to the sensor placement optimization problem and maximize the overall diagnosability. The kernel ideas of the algorithm are described as follows, and the implementation detail of the algorithm can be found in Ozturk et al. (2015).
In the ABC algorithm, the feasible solution to the optimization problem, or a sensor placement x in this paper, is denoted as a food source, with the nectar amount indicating the quality or fitness of the solution. Artificial bees are divided into three categories, that is employed, onlooker and scout bees. Each employed bee is corresponding to a food source, responsible for detecting a new food source in the neighborhood and then memorizing the detected food source greedily when the nectar amount of the detected food source is higher than that of the current one. The genetic operators, including two-point crossover and swap (mutation) operators, are integrated into the binary neighborhood searching procedure. First, two-point crossover operators between xi, xj, xk, xgbest are applied to generate eight children food sources, where xj and xk are two food sources randomly selected from the population of employed bees and xgbest is the best food source among the population. Then, a swap operator is applied to eight children food sources to produce eight grandchildren food sources. The survival food source among the 16 children and grandchildren food sources is selected as the neighborhood food source. After that, employed bees are then sharing information about the food sources with onlookers. Each onlooker selects the food source probabilistically in a process of ‘roulette wheel selection’, based on the nectar amount of each food source. Therefore, quality food sources are selected with higher probability, and thus can be fully exploited. Similarly, each onlooker searches for a new food source in the binary neighborhood of its selected food source and moves to it if the nectar amount of the new food source is higher than that of the selected. If the food source does not improve for a predetermined number of iterations, this food source is deemed to be completely exploited and should be abandoned. These employed bees become scout bees, randomly searching for new food sources in the solution space. In short, exploration and exploitation procedures are carried out together for a robust search for the optimal solution. The flowchart of the enhanced binary ABC algorithm is presented in Figure 1.
RESULTS AND DISCUSSION
This section illustrates the proposed optimal sensor placement strategy with a real-life case study. The network is located in a major city in central China, consisting of two water plants, with 171 nodes and 239 pipes allocated on its distribution mains, shown in Figure 2. In order to acquire an optimal sensor placement under the most general situation, the typical nodal base demands and corresponding diurnal curves are previously estimated and provided. In the remainder of this section, the number of pressure sensors required in the network is roughly estimated by a greedy algorithm, and the enhanced binary ABC algorithm is then adopted aimed at intelligently searching for an optimal sensor placement scheme.
In general, the exhaustive search method, which necessarily provides the exact result, is typically computationally prohibitive. In this paper, the use of the greedy worst-out algorithm is explored to determine the proper number of sensors required for the network shown here. The row of S that would minimize the average mutual coherence maximally is sequentially removed from S at each iteration. Hence, a set of locations that seems not suitable for sensing potential leaks can be acquired in order. The result is shown in Figure 3. As the figure shows, the redundant information on potential leaks is gradually eliminated when the number of sensors decreases from 171 to 52. Mathematically speaking, with the decrease of the number of sensors, the rows in S that are nearly the linear combination of other rows are removed one by one. With a further decrease of the number of sensors from 52, effective information is lost as well. In this case, due to the constraint of sensor costs, the number of deployed sensors is decided to be 26, considering the diagnosability criterion deteriorates only by 1.5% while the costs halve compared to the estimated best result.
The clustering techniques are frequently adopted in the optimal sensor placement searching process to reduce the initial set of candidate sensors, so that the computational complexity can be limited (Sarrate et al. 2014; Blesa et al. 2016). However, ignoring the differences between sensors in one cluster does not meet the fundamental purpose of the optimal sensor placement problem. Alternatively, heuristic methods are introduced to reduce the computational burden. Compared with the genetic algorithm or particle swarm optimization, the enhanced binary ABC algorithm is inherently more suitable in solving binary optimization problems, such as the optimal sensor placement, which is a particular {0, 1} knapsack problem (Ozturk et al. 2015). By using the enhanced binary ABC algorithm, the computation requirement should be affordable unless the scale of WDN is too large.
In this case, the enhanced binary ABC algorithm is proven to be a valid option to reduce the unacceptable computing costs of the exhaustive search (the diagnosability criterion must be evaluated for times). The initial population for the algorithm is generated randomly. As shown in Figure 4, the global minimum average mutual coherence tends to a constant quickly along with the increasing number of generations. A fairly good characteristic of convergence is provided, which needs only about 150 generations to reach a quasi-optimal average mutual coherence of 0.520218. Some fluctuations appear in the evolution of average and current minimum objective function due to the appearance of scout bees. This indicates that the unchanged food sources are discarded in order to avoid being trapped into the local optimal.
The optimal placement of 26 sensors is shown in Figure 5. For branching pipe network, sensors should be deployed near the end of the branch, furthest away from the inlet, while for a loop pipe network, there exists no clear regularity for the optimal sensor placement, and the result heavily depends on both the topology and normal nodal demands.
A reduction in the number of candidate sensor locations does contribute to speeding up the optimization. According to the above analysis result, one-sidedness considering the degree of diagnosability, all candidate sensors that are located at the intermediate section of branching pipes can be directly eliminated, without much influence on the final sensor placement result. In this case, a total number of 29 candidate sensor locations can be eliminated, leaving only 142 candidate locations to be considered.
Index . | Average mutual coherence . | Leak isolability index . |
---|---|---|
Sub-optimal | 0.5206 | 6967.8 |
Quasi-optimal | 0.5202 | 6973.6 |
Index . | Average mutual coherence . | Leak isolability index . |
---|---|---|
Sub-optimal | 0.5206 | 6967.8 |
Quasi-optimal | 0.5202 | 6973.6 |
The proposed method is of guiding significance to deploying pressure instruments in newly built or reformed WDNs. However, in addition to maximizing the degree of diagnosability, the engineering factors are almost equally decisive. Putting the degree of diagnosability aside, in practical applications, the pressure instruments are also preferred to be installed at the points of higher elevation. Because pressure instruments located at higher elevations reduce their requirements for the range of measurements, instruments are most accurate when measuring pressures within 50–75% of the maximum value on the scale. Moreover, the installation and maintenance costs depend on the sensor placement to a large extent as well. In order to provide a good trade-off between the diagnosability, the overall costs of metering, installation and maintenance combined, a node-specific weight can be easily integrated into the sensor placement optimization.
CONCLUSIONS
In this paper, an optimal sensor placement strategy is proposed to enhance the overall performance for model-based leak detection and localization in WDNs. The model-based leak detection and localization technique is based on a relatively small number of pressure measurements and the associated sensitivity-to-leak analysis. By recovering the sparse representation of a leak-induced nodal demand vector, potential leaks can be detected and located. This paper provides an analytical expression for the leak sensitivity matrix, which contributes to calculation efficiency. According to the theory of compressed sensing, the average mutual coherence is adopted to evaluate whether the measurements contain enough information for identifying potential leaks. An optimization method using the enhanced binary ABC algorithm is proposed for optimal sensor placement by minimizing the average mutual coherence, so that the degree of diagnosability can be maximized.
The proposed method is successfully applied to a real-life network located in a major city in central China. By minimizing the average mutual coherence using an enhanced binary ABC algorithm, a quasi-optimal sensor placement can be guaranteed. Based on the resulting optimal sensor placement obtained using the proposed method, the conclusion is reached that, for branching pipes, when candidate sensors located at the intermediate section own similar or lower elevation and similar or higher costs compared with candidate near the end of the branch, it is recommended that those candidates should be eliminated to reduce the number of candidate sensor locations. Otherwise, a node-specific weight should be integrated into the objective function for a balance between maximizing the degree of diagnosability and minimizing the overall costs.
ACKNOWLEDGEMENTS
This work has been funded by the National Natural Science Foundation of China (No. U1509208; 61573313) and by the Key Technology Research and Development Program of Zhejiang Province (No. 2015C03G2010034).