Abstract

A whole building fault (WBF) refers to a fault occurring in one component, but may cause impacts on other components or subsystems, or arise significant impacts on energy consumption and thermal comfort. Conventional methods (such as component level rule-based method or physical model-based method) which targeted at component level fault detection cannot be successfully used to detect a WBF because of the fault propagation among the closely coupled equipment or subsystems. Therefore, a novel data-driven method named weather and schedule-based pattern matching (WPM) and feature-based principal component analysis (FPCA) method for WBF detection is developed. Three processes are established in the WPM-FPCA method to address three main issues in WBF detection. First, a feature selection process is used to pre-select data measurements which represent a whole building’s operation performance under a satisfied status, namely, baseline status. Second, a WPM process is used to locate weather and schedule patterns in the historical baseline database, which are similar to that from the current/incoming operation data, and to generate a WPM baseline. Lastly, real-time PCA models are generated for both the WPM baseline data and the current operation data. Statistic thresholds used to differentiate normal and abnormal (faulty) operations are automatically generated in this PCA modeling process. The PCA models and thresholds are used to detect the WBF. This paper is the first of a two-part study. Performance evaluation of the developed method is conducted using data collected from a real campus building and will be described in the second part of this paper.

1 Introduction

Today, residential and commercial buildings are responsible for more than 40% of the primary energy consumption in the U.S. [1]. Energy wastes are estimated to reach 15–30% of the total building energy consumption due to poorly maintained, degraded, and improperly controlled equipment or systems in commercial buildings in the U.S. [2]. According to the U.S. Department of Energy, it is estimated that faults can increase primary energy consumption in commercial buildings by approximately one quad (11%) of the energy consumed by heating, ventilation, and air conditioning (HVAC), lighting, and larger refrigeration systems [3]. HVAC component or system faults, i.e., malfunctioning control, operation, sensors, and equipment are considered as the top cause for “deficient” building systems [4] and may trigger significantly negative impacts not only on indoor environment and occupants, but also on operational cost, system lifespan, and energy consumption. Automated fault detection and diagnosis (AFDD) tools, which use an automated procedure to discover operational abnormalities and locate root causes in HVAC, lighting, or other building systems [5], can help to reach energy savings by 5–20% in a building [3]. At the same time, the deployment of AFDD tools can also improve indoor environment quality and reduce peak demand and pollution [6].

Numerous investigations have been conducted on the development of AFDD tools in buildings during the past several decades. However, they mainly focus on component or equipment level fault detection and diagnosis such as chiller plants [79], air handling units (AHU) [1012], variable air volume (VAV) terminal units [13], heat pump systems [14], and so on. Reviews for the development of component level AFDD methods and tools can be found in Refs. [2,5,15,16].

Compared with the component level AFDD tool, the development of whole building level AFDD methods, which can detect and diagnose a whole building fault (WBF), has seldom been studied in the past years. In the context of this paper, a WBF is defined as a fault that may occur in one equipment or component but would impact (cause abnormalities) on more than one component and subsystem, or cause significant impacts on the energy consumption or indoor thermal comfort in a building. In a building HVAC system, various components or equipment are closely linked together. The coupled components or subsystems often cause fault propagation [17], i.e., a fault occurring in one component may impact other components or subsystems, triggering additional faults. For example, when an outdoor air damper of an AHU is stuck at a position that is much higher than normal in the cooling season, this fault would not only cause the AHU’s cooling coil valve to be opened at higher than a normal position, but also could cause abnormalities in the primary cooling system, such as a higher chilled water pump speed and/or higher return chilled water temperature. Furthermore, operators or owners of small-sized or medium-sized buildings often only have the time and budget to pay attention to faults that have a whole building level impact, which cannot be completely reflected by a component level AFDD tool. Therefore, it is necessary to develop methods for whole building level diagnostics, which provide a system level view of a building’s operation [18].

In recent years, the development of system level AFDD methods for HVAC systems has been gaining attention. These methods can be divided into two categories: detailed energy model-based methods and data-driven methods. In the detailed energy model-based methods, a system or equipment energy consumption model is first developed. Energy consumption abnormality can be then identified by comparing the real-time energy consumption with the baseline energy model. For example, Bynum et al. [19] developed an Automated Building Commissioning Analysis Tool (ABCAT), which was a simulation tool combining with an expert knowledge-based (i.e., rules summarized based on domain knowledge) system to detect the energy abnormality. The ABCAT used a baseline building energy model which was calibrated by using real energy consumption data collected from the building automation system (BAS), as well as local weather data. Forecasted energy performances from this calibrated simulation model were then compared with the measured data to detect faults. O'Neill et al. [20] developed another energy model-based fault detection and diagnosis (FDD) method. In this approach, energy consumption data from different subsystems including the HVAC system, lighting system, and plug equipment usage were compared with a reference EnergyPlus model, which was used to calculate the expected annual energy consumption to determine if a system had faults. In addition, another top-down strategy by incorporating temporal and spatial partitions was developed to detect HVAC faults across different levels in a building [18]. In this study, building energy consumption flow features were first extracted to detect system faults. Then, energy consumption data were partitioned and grouped to locate fault sources according to system temporal and spatial characteristics.

The advantage of energy model-based AFDD methods is that detailed building energy simulation results, which could provide a relatively accurate energy baseline (if the simulation models are well calibrated) are used to analyze the building’s energy performance. If the developed model could accurately represent subsystems and whole building level operation, fault diagnosis would not be too difficult because both subsystem level and whole building level comparison (i.e., models versus measurements) can be achieved to isolate faults. The main disadvantage of this approach is that developing such a detailed and accurate building energy simulation model is very time-consuming and expensive, and calibrating such models to achieve a subsystem level accuracy for fault isolation remains to be very challenging.

The second category of WBF AFDD method is the data-driven method, which uses BAS interval data and statistical approaches to identify system operating abnormalities and to locate fault root causes. Currently, BASs are increasingly deployed in commercial buildings and a significant amount of operating information from a HVAC system can be efficiently gathered through sensors and BASs. The added data sources bring tremendous benefits to closely monitor the building operating performance. Further, more and more data-driven methods, which exhibit an advantage of cost-effectiveness, high scalability, and easy integration with existing BASs, have been developed in recent years. Some representative data-driven WBF AFDD methods are reviewed below. For example, a statistical process control and Kalman filter-based method was developed for system level fault detection in HVAC systems [21]. Georgescu et al. used spectral methods to decompose diagnostic rule signals generated from trended data from a building management system, into components, which captured the device behavior at temporal and spatial scales [22]. Through this way, fault classifiers had an increased capability to observe abnormal system behaviors. A recursive deterministic perceptron (RDP) neural network was developed to detect and diagnose faults at the entire building level system [23]. In this study, four RDP neural network models were trained to represent four major components (i.e., chiller, coil, fan, and pump) in a HVAC system.

One strategy of using data-driven methods is using the principal component analysis (PCA) method, which is a multivariate statistical method and has been widely used in process monitoring and fault detection [24]. In the building sector, many studies have been conducted on using PCA-based methods for HVAC AFDD. For example, the PCA method was used to detect faults in variable refrigerant flow systems [25], chiller sensors [26], AHUs [27], and VAV terminal units [28]. Furthermore, in order to adapt various operation characteristics in a HVAC system, a variety of modifications have been developed to improve PCA detection accuracy and robustness. For instance, it is found that the detection accuracy and efficiency can be significantly improved when some data preprocessing techniques are used. Hu et al. developed a self-adaptive PCA which removes error samples in the original dataset and improves the method sensitivity [29]. The method was successfully used for chiller sensor fault detection. Li and Wen developed a wavelet transform to preprocess data so that impacts on fault detection performance due to changeable HVAC operations caused by various weather conditions could be minimized [30]. Li et al. developed a PCA-R and support vector data description method to obtain a better fault data distribution and tighter monitoring statistic, so that the fault detection sensitivity can be improved [31].

In this study, two preprocessing techniques, namely, weather and schedule-based pattern matching (WPM) and feature-based PCA (FPCA), have been developed to address three common fault detection challenges in complex HVAC systems, especially at a building level.

The first challenge faced by all HVAC fault detection methods is that different internal and external conditions, such as weather conditions, occupants and other internal loads, have strong influences on HVAC systems’ operation. How to differentiate abnormalities that are triggered by weather/occupancy from those triggered by faults is challenging. More specifically, how to generate a baseline that represents the same HVAC system operation as the incoming data is therefore difficult. Second, a large amount of data are generated from various sensors and are collected by a typical BAS. The high volume of measured data may cause a heavy computation burden for data-driven approaches and may reduce the detection accuracy. Last, a majority of existing methods have been evaluated only through simulation studies. However, data generated through these simulation platforms could hardly reflect a real system’s operational behaviors, especially when a system contains faults. Therefore, it is worthwhile to fully evaluate the developed method by using real BAS data in a building. Besides, using real BAS data for FDD method evaluation would also enable the method to be more easily integrated with an existing BAS.

This paper is the first part of a two-part study. In this paper, the theory of the WPM-FPCA method is introduced in detail. This paper is organized as follows. Section 2 describes the outline of the WPM-FPCA method. Section 3 presents the feature selection process. Section 4 introduces the development of the WPM method. Section 5 illustrated the PCA method for automated fault detection. Finally, Sec. 6 concludes this paper. The performance evaluation of the proposed method is carried out by using BAS data collected from a real campus building and will be described in the second part of this paper.

2 Outline of the Method

The architecture of the WPM-FPCA method consists of three procedures of the detection sequence as shown in Fig. 1.

Fig. 1
Overview of the WPM-FPCA method
Fig. 1
Overview of the WPM-FPCA method
Close modal

The first procedure (Sec. A in Fig. 1) is to perform a feature selection process by using the Partial Least Square Regression and Genetic Algorithm (PLSR-GA) method. A historical baseline dataset would need to be collected first. Here, a baseline is defined as a status at which the building’s operation is considered to be satisfactory, such as when the building has just gone through a commissioning process. For a real building, a true fault-free status is hardly achievable. Hence, a more realistic fault detection process is to detect when a building’s status is significantly different from its baseline status. As discussed in Sec. 1, such a feature selection process, which pre-selects key features from all BAS measurements, is necessary to reduce the dimension of the original BAS dataset. This process is implemented in an offline manner, so that the time spent on real-time fault detection can be reduced. The key features selected from this process are later used to develop the PCA models.

The second procedure (Sec. B in Fig. 1) is to perform a WPM (weather and schedule-based pattern matching). In this study, a symbolic aggregate approximation (SAX, further detailed in Sec. 4.2) algorithm, which is a highly efficient pattern matching (PM) algorithm designed for time series data, is adopted for the WPM process. The SAX algorithm searches the historical baseline dataset by tagging the weather data. Through this tagging process, a WPM baseline dataset, which has similar weather conditions and building use (operation and occupancy) as the incoming snapshot data (current status data collected by the BAS), is generated for each snapshot window.

The third procedure (Sec. C in Fig. 1) is to determine whether a whole building level abnormality exists. The PCA modeling process is used here to evaluate the similarity (T2 value used in this study) between the incoming real-time building operation data and the historical WPM baseline data. If the similarity (i.e., T2 value) between the PCA model from the incoming data and that from the historical WPM baseline data is higher than the threshold (automatically generated when baseline PCA model is developed), the system operation is considered abnormal (faulty). This process can be carried out as an online process.

3 Development of Feature Selection Process

3.1 Introduction of Feature Selection.

While a large amount of HVAC system data (sensor measurements, control signals, etc.) can be easily collected and stored in an advanced BAS, not all of the data are equally important for monitoring system operation. Usually, only very few useful information can be deciphered from this large amount of data. This leads to a “data rich, but information poor” syndrome in the current BAS [30]. Furthermore, the enormous amount of data in building systems may also lower the accuracy of fault detection method [32]. Therefore, a determination and pre-selection of fewer key variables, which can impact the system overall operation and relevant information, should be conducted to improve the detection efficiency and accuracy when implementing fault detection [33].

These key variables can be defined as features of a system in the context of data analytics. Consequently, the determination of these key variables is equivalent to a process of reducing the variable dimensionality. This process can be achieved by optimally designing the monitoring system using expert knowledge, i.e., an expert determines and pre-selects a specific subset of a variable dataset for monitoring system operation [33]. However, in a large-scale system, manually selecting the proper variable subsets and reducing the variable dimensionality become challenging for the following three reasons. First, the number of data measurements has increased rapidly as a result of an increasing complexity of HVAC systems in commercial buildings. Therefore, a manual identification process would take too much time, and would not be suitable for the whole building AFDD solutions which require less human intervention. Second, a system with multi-operation modes may generate various combinations of sub datasets. This makes the expert knowledge-based variables reduction impractical in real practice. Lastly, key variables may often vary from building to building due to a diversity of building design, envelope types, internal load, and system configurations. It causes a manual process for variable selection to be more difficult and time-consuming. Hence, an automated variable dimension reduction technique is determined to be used to implement the variable selection process.

Although the PCA method can partly serve as the dimension reduction function when it is used for fault detection, PCA will be iteratively used to detect the abnormality in a dataset in each snapshot window. In such a detection scenario, a very high data dimensionality will cause a very heavy computational burden. Therefore, the PLSR-GA (explained in Sec. 3.2) method is used to pre-select features which represent the whole building’s operational performance in an offline manner.

3.2 Partial Least Square Regression and Genetic Algorithm Method.

In the PLSR-GA method, PLSR is used to develop a system performance model. Both building electrical power consumption and building virtual cooling consumption from historical baseline dataset can be used as the target output variables (indicators), respectively.

There are many PLSR algorithms for feature selection. Most methods are based on orthogonal score PLSR. The “wrapper” method [34] is one of the most popular methods among them. The PLSR based “wrapper” method is implemented in an iterative way, and is based on some supervised learning approaches, in which a PLSR model refitting is wrapped within the variable searching algorithm. Various variable searching algorithms can be used to obtain a subset of candidate variables and evaluate each subset by fitting a system performance model to the variable subset, so that the model performance can be evaluated and optimized.

Although all possible variable subsets are ideally to be evaluated, it becomes very time-consuming because the number of variable subsets increases dramatically with the number of variables. Therefore, a randomized searching method, i.e., GA [35] is used in the PLSR for feature selection in this study.

3.2.1 PLSR for the Development of a Building Energy Consumption Model.

Partial least square regression is a technique that generalizes and combines features from PCA and multiple regression. Although this technique was not originally designed for classifying and discriminating the dataset, it was found to be efficient to distinguish different datasets and determine the key variables [36].

Let system output Y matrix (YRM×1) consists of M observations and one dependent variable, and system input X (XRM×N) matrix consists of M observations and N dependent variables. The goal of PLS regression is to best predict Y from X and to describe their common structure [37]. Usually, PLSR searches for a set of components (called latent vectors) that decompose simultaneously the X and Y with the constraint that these components maximally explain the covariance between X and Y. This step is similar to the PCA method. By this means, the system input X matrix can be decomposed as
X=TPT+EwithTTT=I
(1)
where TRM×K is the score matrix, PRM×N is the loading matrix, ERM×N is the error matrix, and I is the identity matrix.
The output can be decomposed as
y=Tb+f
(2)
where b = [b1, · · ·, bk]TRK is the loading vector of y, fRM is the error vector.
A nonlinear iterative partial least squares (NIPALS) algorithm is used to develop the PLS model [38]. Suppose that the first latent t1 variable to the (k−1)th latent variables tk−1, the loading vectors p1pk−1, and the loading vector b1bk−1 are given. Then the kth residual input and output can be written as follows:
Xk=Xk1+tkpk1T
(3)
yk=yk1+bktk1
(4)
where tk is a linear combination of the columns of Xk. tk = Xkwk, where wk is the kth weighting vector. It is defined so that the covariance between yk and tk is maximized under ||wk|| = 1.
By using the Lagrange multipliers method, the function to maximize yk and tk can be defined as [39]
Gk=ykTtkμ(||wk||1)=ykTXkwkμ(||wk||1)
(5)
where μ is the Lagrange multiplier. Solving Gkw=0, wk can be derived as
wk=XkTyk||XkTyk||
(6)
The kth loading vector pk and the kth loading bk can be written as
pk=XkTtktkTtk
(7)
bk=ykTtktkTbk
(8)

3.2.2 Genetic Algorithm for Subset Searching.

The GA searching process is used to facilitate the process of searching candidate variable subsets. The GA searching process acts in an iterative way to extract the subset of relevant variables and evaluate each subset by fitting a model with the variable subset. When the system performance criteria are met, the variable subset can be determined. By this means, the key informative variables which represent a system overall performance can be selected, and the variable dimensionality can be reduced accordingly.

The GA is a stochastic searching methodology based on an analogy to natural selection and genetics in biological systems [40]. Compared with other algorithms, the GA has several features: (1) the GA works with a set of candidate solutions known as a population and obtains the optimal solution after a series of iterative computations. This process is similar to the natural selection, in which high-quality solutions to optimization and search problems could be generated through the operators such as mutation, crossover and selection defined in field of biology [41]; (2) the GA uses the objective function information rather than explicit derivative information of the problem set; and (3) the GA can handle large search spaces efficiently, and therefore has a lower risk to obtain a local optimal solution [42]. Based on those characteristics, the GA was originally proposed to solve various single-objective or multi-objective optimization problems when a set of solutions were found to be superior to the rest of the solutions in the search space [43].

In the GA searching algorithm, a given variable subset is represented as a binary string of length n. Here, n is the total number of the available variables and the string can be mimicked as “chromosome” in biology [44]. In the position index i of the string, a zero or one will denote the absence or presence of variable i in the variable set. Through this way, a population of variable subsets (chromosomes) is maintained. The “fitness” is used to determine how likely the variable set can survive and breed into the next generation [45]. A new variable set is created from old variable sets by the process similar to the crossover and mutation. In the process of crossover, two different parent variable sets are mixed together to create an offspring variable subset. In the process of mutation, the indices of a single parent variable set are randomly selected to create an offspring variable subset. The iteration will continue through these two processes until the ending criterion is met. This criterion can be either a finite number of iterations or a certain percentage of the individuals in the population using identical variable subsets.

The procedure of using the GA searching method in PLSR for feature selection is illustrated in Fig. 2. The first step is to generate random variable subsets. In this step, an initial population of variable sets is established by setting bits for each variable randomly, where bit “1” represents selection of corresponding variable while “0” presents nonselection. The approximate size of the variable sets is defined in advance. Then, each individual subset of the selected variables is evaluated. In this step, a PLSR model is fit to each variable set, and the model performance is computed.

Fig. 2
GA method for feature set searching
Fig. 2
GA method for feature set searching
Close modal
After evaluating the variable subsets, the worst half of variables are eliminated. The remaining variables which provide a higher fitness will be retained. Therefore, the population has shrunk to half of its original size. The cross-validation is used to evaluate the model performance. In this study, a contiguous blocks method is used in cross-validation because for time series data and batch data, this method can be convenient for assessing the temporal stability and batch-to-batch stability of a model built from the data. In the contiguous blocks method, each test set is determined by selecting contiguous blocks of n/s objects in the dataset [46]. When determining a variable subset, a model root-mean-square error of cross-validation (RMSECV) is used [47]. The RMSECV provides a “composite” prediction error obtained over all repetitions of the cross-validation procedure. The calculation of RMSECV is given in Eq. (9). The variable subset, which has the lowest RMSECV, is finally selected
RMSECV=i=1n(yiy^i)2n
(9)
where yi is the estimation of y^i and n is the number of cross-validation.

In the third step, a collection of variable sets with higher performance are selected to survive until the next “generation.” In the GA method, crossover and mutation can be used to form new variable sets by (1) crossover of selected variables between the surviving variable sets and (2) changing (mutating) the bit value for each variable by a small probability [48,49].

Finally, a selected variable subset is determined. This subset will be used in the fault detection procedure.

3.3 PLSR-GA Method for Feature Pre-Selection.

In the variable selection process for this study, whole building electricity consumption is used as the predicted variable. All data points collected from a BAS are input variables.

In this study, a medium-sized mixed-use commercial building from Philadelphia, PA, is used for method development and evaluation. More detailed description about this building is provided in Part II of this two-part paper series. In the BAS of this building, 537 data points (measurements, control signals, etc.) are collected at a 5 min interval. Historical baseline datasets have been collected for summer, winter, and transitional seasons. For this building, there are three operation mode, i.e., Mode #1 and Mode #2 for summer season, as well as Mode #3 for transitional and winter seasons. Each mode here refers to a different operation strategy/schedule. More details are again provided in Part II of the study.

For each operation mode, the feature selection procedure is divided into several selection scenarios. This is because one of the risks in the GA variable selection is overfitting. Therefore, repeating the GA for several times and obtaining general trends can help to lower the risk of overfitting [47]. If a variable is repeatedly selected from different scenarios, this variable is more likely to represent the system operational performance for the entire operation mode.

In this research, every five consecutive days are grouped into one selection scenario in each operation mode. If a variable is selected in three scenarios, then the variable will be included in the final feature list to be used in the development of the PCA model.

3.4 Evaluation of PLSR-GA Feature Selection.

Although, Akaike Information Criterion [50] and Bayesian Information Criterion [51] can be used to evaluate the feature selection efficacy by analyzing the model developed from the selected variables, the ending purpose is not to evaluate the performance of the regression model developed from the selected variable set. Therefore, the results were evaluated by manually examining the selected variables.

In this study, the evaluation of the PLSR-GA feature selection is divided into two processes. In the first evaluation process, the criterion is established based on the empirical method to examine the feature selection ratio on the key measurements from the expert’s knowledge. This process is common in the field of image process and process control industry. In the second evaluation process, the criterion is to use the fault detection result to demonstrate the effectiveness of feature selection.

For a building system, which has many coupled subsystems, it is quite challenging to choose one or two measurements or control signals as benchmarks to assess an overall system operation. In the study, a total of 30 measurements from the HVAC system are manually selected as key measurements to evaluate the feature selection result. Among these measurements, measurements No. 1 to No 9. are from the chiller plant and are only used during a summer season (i.e., operation mode 1 and mode 2). Measurements No. 10 to No. 30 are from AHU-1 to AHU-3. The outdoor air damper position of AHU-3 is not included as a key measurement because the outdoor air damper in AHU-3 is opened at 100% position during all seasons as the AHU-3 serves lab areas. The steam valve positions for AHU-1 to AHU-3 (measurements Nos. 28, 29, and 30) are only used in the operation mode in transitional and winter seasons. The cooling coil valve positions for AHU-1 and AHU-3 (measurements No. 13, 19, and 25) are only used in evaluating variable selection result in summer season (i.e., operation mode #1 and #2). Weather conditions (outdoor air temperature, outdoor air humidity and outdoor air enthalpy) are not included here because those measurements are not used to develop the PCA model.

Therefore, a total of 27 measurements are used in the evaluation for operation mode #1 and operation mode #2. And a total of 18 measurements are used in the evaluation for operation mode #3. All key measurements in each operation mode are provided in Table 1. The word “variable” instead of “measurement” will be used in the following context to illustrate the feature selection results.

Table 1

Key measurements (variables) from the BAS

No.Name of measurementMode#1Mode#2Mode#3No.Name of measurementMode#1Mode#2Mode#3
1CHWS temperature×16AHU-2 Supply air temperature
2CHWR temperature×17AHU-2 Supply air static pressure
3Cooling water supply temperature×18AHU-2 Supply air fan speed
4Cooling water return temperature×19AHU-2 Cooling coil valve position×
5Chilled water differential pressure×20AHU-2 Outdoor air damper position
6Chilled water pump speed×21AHU-2 Mixed air temperature
7Chilled water flowrate×22AHU-3 Supply air temperature
8Cooling water differential pressure×23AHU-3 Supply air static pressure
9Cooling water pump speed×24AHU-3 Supply air fan speed
10AHU-1 Supply air temperature25AHU-3 Cooling coil valve position×
11AHU-1 Supply air static pressure26AHU-3 Supply air humidity
12AHU-1 Supply air fan speed27AHU-3 Between-coil air temperature
13AHU-1 Cooling coil valve position×28AHU-1 Steam valve position××
14AHU-1 Outdoor air damper position29AHU-2 Steam valve position××
15AHU-1 Mixed air temperature30AHU-3 Steam valve position××
No.Name of measurementMode#1Mode#2Mode#3No.Name of measurementMode#1Mode#2Mode#3
1CHWS temperature×16AHU-2 Supply air temperature
2CHWR temperature×17AHU-2 Supply air static pressure
3Cooling water supply temperature×18AHU-2 Supply air fan speed
4Cooling water return temperature×19AHU-2 Cooling coil valve position×
5Chilled water differential pressure×20AHU-2 Outdoor air damper position
6Chilled water pump speed×21AHU-2 Mixed air temperature
7Chilled water flowrate×22AHU-3 Supply air temperature
8Cooling water differential pressure×23AHU-3 Supply air static pressure
9Cooling water pump speed×24AHU-3 Supply air fan speed
10AHU-1 Supply air temperature25AHU-3 Cooling coil valve position×
11AHU-1 Supply air static pressure26AHU-3 Supply air humidity
12AHU-1 Supply air fan speed27AHU-3 Between-coil air temperature
13AHU-1 Cooling coil valve position×28AHU-1 Steam valve position××
14AHU-1 Outdoor air damper position29AHU-2 Steam valve position××
15AHU-1 Mixed air temperature30AHU-3 Steam valve position××
The whole variable selection ratio (WVSR) was defined to compare the result as given below:
WVSR=NSVNWV%
(10)
where NSV is the number of selected variables, NWV is the number of whole variables.
In order to evaluate the key variable selection result, the key variable selection ratio (KVSR) was defined as given below:
KVSR=NSKVNKV%
(11)
where NSKV is the number of selected key variables, NKV is the number of predefined whole key variables. Here, the NKV is 27 for modes #1 and #2, and 18 for mode #3.
At the same time, the key variable selection efficiency ratio (KVSER) was defined to evaluate whether the variable selection method has high efficiency to automatically identify key variables from all the manually selected system variables. KVSER is given below:
KVSER=NSKVNSV%
(12)

From this equation, a higher NSKV or a lower NSV will lead to a higher KVSER. This means that the variable selection method can actually better identify a key variable after the variable selection process.

3.4.1 Feature Selection Result Under the Operation Mode #1.

In mode #1, 40 historical baseline days are included. Eight variable selection scenarios are implemented.

A total of 101 variables are selected in the end. Among them, 12 key variables are included as given in Table 1. In operation mode #1, the chiller provides a relatively stable CHWS temperature, and the chilled water flowrate is also stable during all time in a day. Therefore, less variables from the chiller plant are selected. A total of 80 variables are from VAV terminals in different zones. Table 2 shows the selection analysis. Table 3 shows the selected key variables.

Table 2

Variable selection analysis (mode #1)

Number of selected variablesNumber of selected key variablesWVSRKVSRKVSER
1011218.8%44.4%11.9%
Number of selected variablesNumber of selected key variablesWVSRKVSRKVSER
1011218.8%44.4%11.9%
Table 3

Selected key variable list (mode #1)

No.Selected key variable nameNo.Selected key variable name
1Cooling water return temperature7AHU-2 Supply air temperature
2CHWS delta pressure8AHU-2 Supply air fan speed
3Cooling water delta pressure9AHU-2 Cooling coil valve position
4AHU-1 Supply air fan speed10AHU-2 supply air static pressure
5AHU-1 Cooling coil valve position11AHU-3 Supply air temperature
6AHU-1 Supply air static pressure12AHU-3 Between-coil temperature
No.Selected key variable nameNo.Selected key variable name
1Cooling water return temperature7AHU-2 Supply air temperature
2CHWS delta pressure8AHU-2 Supply air fan speed
3Cooling water delta pressure9AHU-2 Cooling coil valve position
4AHU-1 Supply air fan speed10AHU-2 supply air static pressure
5AHU-1 Cooling coil valve position11AHU-3 Supply air temperature
6AHU-1 Supply air static pressure12AHU-3 Between-coil temperature

3.4.2 Feature Selection Result Under the Operation Mode #2.

In mode #2, 50 historical baseline days are included. Ten variable selection scenarios are implemented.

Two hundred and fourteen variables are selected in the end. Among them, there are 11 key variables. In operation mode #2, two variables from weather conditions including outdoor air temperature and enthalpy are selected. One hundred and nineteen variables are from VAV terminals in different zones. Table 4 shows the selection analysis. Table 5 shows the selected key variables.

Table 4

Variable selection analysis (mode #2)

Number of selected variablesNumber of selected key variablesWVSRKVSRKVSER
2141139.9%40.7%5.1%
Number of selected variablesNumber of selected key variablesWVSRKVSRKVSER
2141139.9%40.7%5.1%
Table 5

Selected key variable list (mode #2)

No.Selected key variable namesNo.Selected key variable names
1Chilled water flowrate7AHU-2 Supply air static pressure
2Cooling water supply temperature8AHU-3 Supply air temperature
3Cooling water delta pressure9AHU-3 Supply air humidity
4AHU-1 Cooling coil valve position10AHU-3 Supply air fan speed
5AHU-1 Supply air static pressure11AHU-3 Between-coil temperature
6AHU-2 Cooling coil valve position
No.Selected key variable namesNo.Selected key variable names
1Chilled water flowrate7AHU-2 Supply air static pressure
2Cooling water supply temperature8AHU-3 Supply air temperature
3Cooling water delta pressure9AHU-3 Supply air humidity
4AHU-1 Cooling coil valve position10AHU-3 Supply air fan speed
5AHU-1 Supply air static pressure11AHU-3 Between-coil temperature
6AHU-2 Cooling coil valve position

3.4.3 Feature Selection Result Under the Operation Mode #3.

In mode #3, 50 historical baseline days are included. Ten variable selection scenarios are implemented.

A total of 171 variables are selected. Among them, there are six key variables. One hundred and forty-five variables are from VAV terminals in different zones. Table 6 shows the selection analysis. Table 7 shows the selected key variables.

Table 6

Variable selection analysis (mode #3)

Number of selected variablesNumber of selected key variablesWVSRKVSRKVSER
171631.8%33.3%3.5%
Number of selected variablesNumber of selected key variablesWVSRKVSRKVSER
171631.8%33.3%3.5%
Table 7

Selected key variable list (mode #3)

No.Selected key variable namesNo.Selected key variable names
1AHU-1 Supply air fan speed4AHU-2 Steam valve position
2AHU-1 Steam valve position5AHU-3 Supply air humidity
3AHU-2 Supply air static pressure6AHU-3 Between-coil temperature
No.Selected key variable namesNo.Selected key variable names
1AHU-1 Supply air fan speed4AHU-2 Steam valve position
2AHU-1 Steam valve position5AHU-3 Supply air humidity
3AHU-2 Supply air static pressure6AHU-3 Between-coil temperature

3.4.4 Discussion.

The feature selection process successfully lowers the data dimensionality in all three operation modes. The highest key variable selection rate and key variable selection efficiency are achieved for Operation mode #1. Despite that a higher WVSR is achieved in operation modes #2 and #3, the KVSERs for both modes are not as high as compared with mode #1. On the contrary, more variables from the VAV terminals are selected in those two operation modes (i.e., modes #2 and #3). Therefore, the evaluation of an overall performance of a HVAC system may include more downstream subsystems if the chiller’s operation is not stable or the chiller is not operated.

Although it is recommended that the first procedure in implementing a feature selection is to use the domain knowledge [52], in this research, expert knowledge was not included to modify the selection results. Furthermore, during the feature selection process, only the PLSR-GA method was used to select features that reflect a satisfied system’s operation. Therefore, it may cause some limitations such as key measurements that do not affect the system’s performance during a normal operation may not be selected. For example, the CHWS temperature is maintained to be a relative stable value (i.e., 6.7 °C in modes #1 and #2) under normal operation. Therefore, although the measurement of CHWS temperature can be very useful in judging a chiller’s operation status, it is not selected and not used in the fault detection process.

Based on the limitations above, the above-described feature selection result is recommended to only be used in the fault detection process (when one needs to judge the abnormality from normal operation), but not in the fault diagnosis process (when one needs to isolate the root cause of an abnormality and hence needs features that reflect system operation with faults).

4 Development of Weather and Schedule-Based Pattern Matching Method

4.1 Introduction of Pattern Matching.

In a HVAC system, different internal and external conditions such as weather, occupancy schedule, and usages of electrical appliances have direct influences on the system’s operation. For example, when outdoor air warms up in the morning hours, the air handling unit could change its operation from a heating mode, to an economizer mode and then to a cooling mode. In such cases, directly using the data during a long span of time from the historical dataset may generate inaccurate baseline and make the fault detection process difficult as the abnormality could be caused by faults or by other internal/external conditions.

To address this issue, a PM method was developed to locate periods of system operation that are similar to the current operating conditions. This method was successfully used in the AHU fault detection [10]. When developing a PM method, two aspects, which have impacts on the thermal load of a building, and further have impact on a HVAC system operation, need to be considered. One aspect is those external factors that have significant impacts on the thermal load such as the weather conditions. Another aspect is those factors that affect a building’s internal thermal load, such as building occupancy and the usage of electrical appliances. Under the same operation mode, the operation behavior of a fault-free HVAC system is similar to that from the historic baseline.

Finding and matching similar patterns of internal and external factors to generate operation baseline data that has a similar operation as that from the incoming data are hence important to differentiate abnormalities caused by faults from those caused by other factors. This section discusses methods developed for matching weather patterns. Notice that similar methods can be used to match factors related to internal load, if such measurements are available (i.e., occupancy and appliance usages etc.). For the building considered in this study, such measurements related to internal loads are not available. However, these internal loads are highly correlated to the time of the day. Hence, time of the day (schedule) is directly used to represent the internal load factors in this study. This situation, i.e., lacking measurements related to internal load, is not uncommon for most of the commercial buildings.

4.2 Symbolic Aggregate Approximation Method.

Because the weather information data are a time series data, techniques for time series data pattern matching are considered. In this research, the SAX time series method is used to search similar weather time series data from the historical database. The SAX method is a piecewise aggregate approximation (PPA)-based method to identify similarity of time series data [53]. The PPA uses an equal-sized segment time series data cutting technique. In the PPA, a set of time series which constitutes the database as Y = {Y1, · · ·, Yk} is compared with the original time series dataset X = {x1, · · ·, xk}. Each sequence in X is n units long. Let N be the dimensionality of the transformed space and set the index (1 ≤ Nn). Here, N is assumed as a factor of n [54]. Then, the original time series data X of length n can be represented in N space by a vector X¯=x¯1,,x¯k. The ith element of X¯ can be calculated by
x¯i=Nnj=nN(i1)+1(nN)ixj
(13)

By this means, a time series data is divided into N equal-sized segments. At the same time, the data with n dimensions can be reduced to N dimensions. Then, the mean value of the data in each segment can be calculated. Therefore, this transformation produces a piecewise constant approximation of the original time series dataset. Figure 3 illustrates this transformation based on the PPA method. The x axis represents the time series and the y axis represents the data range.

Fig. 3
Illustration of PPA method
Fig. 3
Illustration of PPA method
Close modal

4.3 Development of Weather and Schedule-Based Pattern Matching Method.

When using the SAX method, the historical dataset should be first generated. Then, the weather information data from the historical dataset and the new incoming snapshot data are combined together to produce one time series dataset. The weather information of incoming snapshot data is extracted from this time series dataset to generate X(t). The weather information data X(t) is normalized by using z-score normalization.

After normalization, weather information data X(t) is divided into N individual nonoverlapping segments which are equal-sized data segments. One data segment is defined as one window. In this study, the same window size is applied for both snapshot data and historical baseline dataset. The window size is determined by two factors: how fast weather conditions change and computation efficiency. If the window size is too large, weather changes largely within a data window which leads to a significant operation change of the system. This will reduce the PM accuracy. If the window size is too small, within each data window, only a small number of data samples will be included, which may not be sufficient for accurate comparison. Moreover, the computation burden for fault detection and diagnosis will be increased. How this window size is selected is discussed in more detail in Sec. 4.4.

A number of alphabets are used to convert each window into a SAX string or letter. Each alphabet corresponds to a subsequence of data. In this study, the number of alphabets is set to 10 according to the SAX [55]. In the study, the outdoor air enthalpy is used as the tagging target because air enthalpy includes both temperature information and humidity information.

When the weather information time series data is tagged, the system’s operation data segments with the same symbolic letters/strings are clustered to generate a WPM baseline dataset. In order to control sample size, in this research, a sample pool size S is used to define how many samples are used to generate a WPM baseline dataset. As mentioned above, although a smaller window size can more accurately capture the weather change, less information is obtained to generate a valid baseline database. Therefore, an adjacent window (referred as data sample searching pool) to the snapshot window can be used as the sample pool because the building occupancy may remain steady in a relatively long period. Figure 4 summarizes the above-discussed procedure of WPM by using the SAX method. Details about how to select both window size and sample pool size is provided in Sec. 4.4.

Fig. 4

4.4 Selection of Weather and Schedule-Based Pattern Matching Parameters

4.4.1 Introduction.

When implementing the SAX method for the WPM, different parameter settings such as the number of baseline days, snapshot window size, and searching pool size, may have impacts on the number of samples in each snapshot window, as well as the WPM accuracy. Therefore, it is necessary to determine a proper parameter setting to ensure (1) the number of samples in each snapshot window, so that the later data-driven method can be implemented, and (2) the PM accuracy, so that better detection results can be achieved. Sensitivity analyses are carried out to include several test cases under different parameter settings as illustrated in Table 8. Three scenarios are conducted as

  1. Number of WBF-free days: 20 days, 30 days, and 40 days. Notice that here WBF-free days refer to the baseline days, in which the building is considered to be without any WBF. The baseline is needed for a data-driven fault detection process. However, the more baseline days a fault detection strategy requires, the more challenging it is to collect the required data.

  2. Snapshot window size: 15 min, 30 min, and 45 min.

  3. Searching pool size: 120 min, 180 min, and 240 min.

Table 8

Test description

Test nameTest descriptionTest cases
Number of WBF-free dayTest the minimum number of WBF-free day that is needed to achieve a satisfactory performance20-day, 30-day, and 40-day
Snapshot window sizeTest the impacts of the snapshot window size on the samples selected in each snapshot window15-min, 30-min, 45-min
Searching pool sizeTest the impacts of the searching pool size on the samples selected in each snapshot window120-min, 180-min, 240-min
Test nameTest descriptionTest cases
Number of WBF-free dayTest the minimum number of WBF-free day that is needed to achieve a satisfactory performance20-day, 30-day, and 40-day
Snapshot window sizeTest the impacts of the snapshot window size on the samples selected in each snapshot window15-min, 30-min, 45-min
Searching pool sizeTest the impacts of the searching pool size on the samples selected in each snapshot window120-min, 180-min, 240-min

The testing matrix on the parameters will become very complicated due to large combinations. Therefore, in this study, a baseline case is first defined. In this baseline case, the parameter selections are number of WBF-free day equals 30 day, snapshot window size equals 30 min, and sample pool size equals 180 min.

4.4.2 Historical Datasets and Evaluation Criterion.

Since the WPM method assumes that a building is operated under similar operating modes when it is subject to similar weather conditions (which is a typical situation), if a building has multiple operating modes under a similar weather condition, each operating mode requires a corresponding historical baseline. This is not very typical but has occurred in the demonstration building in this study. The facility has modified the building’s operation strategies between 2016 and 2017. Hence, a historical dataset for each year is constructed for each season. In this research, the test of the WPM method is implemented within the same season.

Here seven historical datasets from 2016 to 2017, i.e., spring 2017, summer 2016 and 2017, fall 2016 and 2017, as well as winter 2016 and 2017, are used to analyze the impact of different parameter settings on the WPM performance. In each dataset, 10 days are randomly selected as one test case to obtain the WPM baseline dataset.

Four quantitative criteria are used to evaluate the WPM performance, i.e., (1) maximum number of sample (MaxNS) in each snapshot window, (2) minimum number of sample (MinNS) in each snapshot window, (3) average number of sample (AveNS) in each snapshot window, and (4) percentage of snapshot window (PctWin) with sample size less than 100.

The AveNS is defined below:
AveNS=i=1nNS/n
(14)
where NS is the total number of samples, n is the number of snapshot windows used in each test case. For example, for 10 test days under 30 min snapshot window size, the total number of snapshot windows is equal to 480.

It is noted that there is a lack of recommendation from the existing literature on the minimum data samples that should be included in each snapshot window. Usually in a BAS, the sampling rate ranges from 1 min to 15 min. Here, the data sampling rate is 5 min in the BAS of the tested building. Time range of data collection for a building in one season can be from 20 days to 60 days depending on the actual operation. Therefore, for a 20-day to 60-day historical baseline day, the minimum and maximum number of data samples in a 30-min snapshot window can be 120–300 under 100% selection rate. Here the AveNS is set to be equal to 100 as a threshold, i.e., the parameters, which ensure the AveNS more than 100 in the WPM method is used.

For the minimum number of historical baselines based on the existing literature and some preliminary tests, 20 days, 30 days, and 40 days are selected to be tested. When this parameter is being tested, other parameters are set to be their default values, i.e., snapshot window size equals 30 min, and sample pool size equals 180 min.

4.4.3 Selection of Number of Historical Baseline Days.

Obviously, including more historical baseline days, i.e., WBF-free days, will increase the accuracy and robustness of the fault detection method. However, in the real practice, especially for building re-commissioning, it may not always be possible to obtain a large number of historical WBF-free days. Therefore, it will be meaningful to investigate a suitable number of WBF-free days used to generate a baseline.

In this research, 20-day, 30-day, and 40-day historical baseline days are used in the sensitivity analysis, respectively. Seven test cases corresponding to different seasons are collected from the test building. The sizes of historical baseline days for each test case are from 61 days to 63 days as given in Table 9.

Table 9

Selection of minimum number of historical baseline days

Test casesSize of historical baseline (days)MaxNSMinNSAveNSPctWin
203040203040203040203040
Fall 201661180216348112308212617062%22%13%
Fall 2017622043124321118413323066%33%14%
Spring 2017622823184501119514519359%32%27%
Summer 201662288342438113012217021840%14%9%
Summer 201761210264354130307212217575%44%28%
Winter 20166328841447416610415919543%31%27%
Winter 201762258324348130309114518753%26%17%
Test casesSize of historical baseline (days)MaxNSMinNSAveNSPctWin
203040203040203040203040
Fall 201661180216348112308212617062%22%13%
Fall 2017622043124321118413323066%33%14%
Spring 2017622823184501119514519359%32%27%
Summer 201662288342438113012217021840%14%9%
Summer 201761210264354130307212217575%44%28%
Winter 20166328841447416610415919543%31%27%
Winter 201762258324348130309114518753%26%17%

The function of a historical baseline is to provide baseline data for incoming snapshot data. In this study, whenever a snapshot data comes in (for example 10:00–10:30, outdoor temperature is around 26.7 °C and relative humidity is around 70%), data from baseline dataset that are within the same time (10:00–10:30) and with similar weather (26.7 °C and 70% relative humidity) will be collected to form the WPM baseline for this specific snapshot data. Hence, the key of minimum historical baseline data collection really depends on how repeatable the weather is at a specific location.

In this study, for each test case, 10 days are randomly selected and considered as the incoming snapshot data. The sensitivity test will then randomly pick the remaining baseline days to form the historical baseline dataset with varying sizes.

Table 9 shows the test result. Again, four indices defined earlier are used here to evaluate the impact of varying historical baseline size on the number of samples that are selected to form the WPM baseline. The results show that, only in the test cases of summer and winter 2016, the AveNS can surpass 100 when using 20 historical days to generate baseline datasets. At the same time the PctWin for all test cases is more than 40%. This means that more than 40% of the window in the baseline has data samples less than 100. Therefore, 20-day is not used for the WPM method to generate the WPM baseline dataset.

When using 30 historical days to generate a baseline dataset, the AveNS covers from 122 to 170 in all seven test cases. At the same time, the PctWin ranges from 14% to 44%. When using 40 historical days to generate a baseline dataset, the AveNS covers from 170 to 230 in all seven test cases. At the same time, the PctWin ranges from 9% to 28%. Despite 40-day historical baseline generates better results in both AveNS and PctWin, considering in the real practice, the collection of more historical days becomes increasingly difficult, 30 days are selected as the minimum number of historical baseline days in this study.

4.4.4 Selection of Snapshot Window Size.

In order to identify the effects of snapshot window size on the WPM method and fault detection result, sensitivity tests are performed. A dataset with smaller window size helps the SAX method to better capture the weather dynamic change, and therefore generate more accurate baseline data. This will further help to detect faults with higher dynamic responses under different weather and indoor conditions.

In this study, snapshot window sizes of 15 min, 30 min, and 45 min are used to evaluate their impact on the number of data samples in each snapshot window. Data from the seven datasets are used to implement the sensitivity test. Historical datasets that cover a 30 WBF-free day range are used in each test. The criterion of AveNS and PctWin is also used to quantify the sensitivity test result.

Although a smaller window size (15 min or 30 min) can generate a more accurate WPM baseline from the historical baseline dataset than a larger (45 min) window size, too small a snapshot window size may cause a high computation burden.

Besides, the test results show that there is no significant impact on the number of data samples between 15-min and 30-min snapshot window sizes as shown in Table 10. Based on these analyses above, a window size of 30 min will be adopted in the WPM method.

Table 10

Sensitivity test on snapshot window size

Snapshot window size15-min30-min45-min
AveNSPctWinAveNSPctWinAveNSPctWin
Summer 201617214%17014%17017%
Fall 201612520%12622%12921%
Winter 201615931%15931%16131%
Spring 201714631%14532%14733%
Summer 201712244%12244%12246%
Fall 201713333%13333%12740%
Winter 201714427%14526%14628%
Snapshot window size15-min30-min45-min
AveNSPctWinAveNSPctWinAveNSPctWin
Summer 201617214%17014%17017%
Fall 201612520%12622%12921%
Winter 201615931%15931%16131%
Spring 201714631%14532%14733%
Summer 201712244%12244%12246%
Fall 201713333%13333%12740%
Winter 201714427%14526%14628%

4.4.5 Sensitivity Test on Data Sample Searching Pool Size.

When forming the “similar baseline data,” neighboring windows of the snapshot window are typically included in the searching pool for the WPM baseline. For example, when analyzing a snapshot window of 10:00 and 10:30, all collected historical baseline data between 10:00 and 10:30, as well as its neighboring windows, such as 9:30 to 10:00 and 10:30 to 11:00, can be used to form the searching pool, among which, those with similar weather conditions will be selected as the “similar baseline data,” as shown in Fig. 5. This searching pool size (S) will affect the WPM accuracy. If S is too small, there would not be enough samples to be included in the WPM baseline dataset. If S is too large, the time frame of the neighboring windows will be too long and the assumption of invariant internal load may not be correct. In the above example, when extending the concept of neighboring windows to be from 8:00 to 12:00, instead of 9:00 to 11:00, data that contain very different internal loads could be included in the searching pool, and thus affect the WPM accuracy.

Fig. 5
Illustration of data searching pool
Fig. 5
Illustration of data searching pool
Close modal

In order to test the data sample searching pool size, three searching sizes such that 120 min (2 windows before and 2 windows after), 180 min (3 windows before and 3 windows after), and 240 min (4 windows before and 4 windows after) are considered. Here, the snapshot window size is chosen as 30 min, based on the previous sensitivity test. The testing data used here are similar to those in the evaluation of the historical baseline dataset size. Again, when evaluating the data sample searching pool size, AveNS and PctWin are used to evaluate the WPM result under different data sample searching pool sizes. In order to fully evaluate the effects of different data sample searching pool sizes, three numbers of search days, i.e., 20 days, 30 days, and 40 days are used here.

Table 11 summarizes the test results for various data sample searching pool sizes in different datasets. The results show that under a 120 min data sample searching pool size, AveNS can meet the requirement of 100 minimum data samples only when the searching day reaches 40 days. But for the 180-min data sample searching pool size, both 30 searching days and 40 searching days can meet the requirement. Although the requirement can be met for most test cases under a 240-min data sample searching pool size, it may lower the WPM accuracy. Therefore, considering the trade-off of the minimum data sample in each snapshot window, the searching day as well as the searching pool size, the 180-min searching pool size will be used in the WPM method.

Table 11

Sensitivity test on data searching pool size

Searching pool size120-min Searching pool180-min Searching pool240-min Searching pool
MaxNSMinNSAveNSPctWinMaxNSMinNSAveNSPctWinMaxNSMinNSAveNSPctWin
20 searching days
Su 201620118463%288112240%366115928%
F 201612615591%18018262%240110943%
W201616816982%288110443%378113732%
Sp 201720417084%28219559%378112246%
Su 201715015190%21017275%29419365%
F 201714415679%20418466%264110652%
W 201717415984%25819153%330112141%
30 searching days
Su 2016240111541%342117014%438122711%
F 201617468466%2161212622%2822416814%
W2016276110647%414615931%5401221124%
Sp 2017228110056%318114532%426119224%
Su 2017180128264%2643012244%3602416034%
F 201721018956%312113333%414116336%
W 2017228249858%3243014526%4263019319%
40 searching days
Su 20162943014730%438302189%558422916%
F 20162401811434%3483017013%456302279%
W2016312113134%474619524%6181225920%
Sp 2017318113035%450119327%576125717%
Su 20172461811644%3543017528%4383022818%
F 2017288115319%432123014%582130613%
W 2017288115319%3483018717%4503024816%
Searching pool size120-min Searching pool180-min Searching pool240-min Searching pool
MaxNSMinNSAveNSPctWinMaxNSMinNSAveNSPctWinMaxNSMinNSAveNSPctWin
20 searching days
Su 201620118463%288112240%366115928%
F 201612615591%18018262%240110943%
W201616816982%288110443%378113732%
Sp 201720417084%28219559%378112246%
Su 201715015190%21017275%29419365%
F 201714415679%20418466%264110652%
W 201717415984%25819153%330112141%
30 searching days
Su 2016240111541%342117014%438122711%
F 201617468466%2161212622%2822416814%
W2016276110647%414615931%5401221124%
Sp 2017228110056%318114532%426119224%
Su 2017180128264%2643012244%3602416034%
F 201721018956%312113333%414116336%
W 2017228249858%3243014526%4263019319%
40 searching days
Su 20162943014730%438302189%558422916%
F 20162401811434%3483017013%456302279%
W2016312113134%474619524%6181225920%
Sp 2017318113035%450119327%576125717%
Su 20172461811644%3543017528%4383022818%
F 2017288115319%432123014%582130613%
W 2017288115319%3483018717%4503024816%

Note: Su: summer, F: fall, W: winter, and Sp: spring.

5 Principal Component Analysis for Fault Detection

5.1 Introduction of Principal Component Analysis.

The last procedure of the developed method is to use the WPM baseline dataset to develop a statistical model and perform the fault detection over the incoming snapshot data.

In this research, PCA was used for fault detection. PCA is one of the most popular data-driven methods used in various industries to monitor system operation. PCA uses orthogonal transformation procedure to extract a set of linearly uncorrelated principal components (PCs) from the possibly correlated original variables [56]. In PCA, a data measurements matrix XRn×m includes n samples and m process variables. X can be decomposed into a principal matrix X^ plus a residual matrix E. The principal matrix X^ is a product of a score’s matrix T and a loadings matrix P, as given as below:
X=TPT+E=i=1atipiT+E
(15)
where ti is a score vector (orthogonal) which contains information about relationship between samples and pi is a loading vector (orthonormal) which contains information about the relationship between variables. a is the number of PCs which are selected to retain in the model.
Although, there is not a best strategy to determine how many PCs are retained in the model, rules of thumb from the literature and user’s own experiences are typically used to determine the number of PCs [47]. In this study, the cumulative variance percentage contribution method as provided in Eq. (16) is used. In this method, the smallest number of PCs is selected to capture a certain percentage of the cumulative variance which is predetermined by the users. This number of PCs contributes the maximum cumulative variance [57]
d1+d2++dqd1+d2++dpθ
(16)
where θ is the predetermined percentage, d is variance captured by each PC, p is the total number of PCs, and q is the retained number of PC. For example, if θ is set to 0.95, 95% cumulative variance will be retained in the PCA model as shown in Fig. 6. In Fig. 6, 11 PCs which represent 96.1% cumulative variance are retained in the PCA model. In this study, different PC retention rates are studied to test detection results.
Fig. 6
Scree plot example of cumulative variance percentage contribution method
Fig. 6
Scree plot example of cumulative variance percentage contribution method
Close modal
In the study, T2 statistic is used to obtain the detection result and can be calculated for each new observation by
T2=xTPaPTx
(17)
where a contains the non-negative eigenvalues corresponding to a PCs, and x is a newly obtained process data vector.
The upper confidence threshold of T2 can be calculated by using F-distribution:
Tthreshold2=a(n1)naFa,na,α
(18)
where n is the number of samples in the data, α is the level of significance. In this study, α is set to 0.01.

5.2 Workflow of Fault Detection.

Figure 7 demonstrates the workflow of PCA fault detection for a single snapshot window. When implementing the PCA method for fault detection, the following procedures are followed. First, a “WPM baseline dataset,” i.e., data that have similar weather condition and time frame as the snapshot data, is identified from the historical dataset by using the WPM method. Second, a local PCA model is developed for the “WPM baseline dataset” of each snapshot window. Then, a statistical threshold (i.e., T2) is generated for the developed local PCA model in each snapshot window. Third, the statistical value from the snapshot data will be calculated and compared with the threshold. Lastly, abnormal snapshot data is flagged and marked if the statistical value (from incoming snapshot data) is higher than the statistical threshold. The same procedure will be repeated when a new incoming snapshot data is collected and fed into the fault detection process. Through this procedure, a dynamic characteristic of a system operation can be more accurately modeled. Therefore, the detection accuracy is expected to be increased.

Fig. 7
Flowchart of WPM-FPCA for fault detection
Fig. 7
Flowchart of WPM-FPCA for fault detection
Close modal

6 Conclusions

In this first paper of a two-paper series, the principles and methodology of a novel data-driven fault detection and diagnosis method, i.e., the WPM-FPCA method, is presented for detecting whole building level faults. In this method, a feature pre-selection process is first implemented by using the PLSR-GA method to select candidate variables, which are later used to develop PCA models. The HVAC system operation schedule as well as weather patterns, such as outdoor air enthalpy, are used to dynamically generate baseline data (i.e., WPM baseline data) that represent a building’s normal operation. A real-time PCA model is continuously generated based on the selected features for each WPM baseline according to each incoming data window. When a new incoming BAS data is collected and fed into the fault detection solution, system operational abnormality can be flagged if the calculated statistics T2 overpasses the threshold which is automatically generated from the baseline PCA model. The selections of several parameters such as the number of baseline days, snapshot window size, and data sample searching pool are also discussed. The implementation process of the WPM-FPCA method to detect WBFs in a real building, as well as the evaluation of the developed WPM-FPCA method by using real BAS data will be presented in the second part of this study.

Acknowledgment

Financial support provided by the U.S. Department of Energy for the research of VOLTTRON Compatible Whole Building Root-Fault Detection and Diagnosis (Grant No. DE-FOA-0001167) is greatly appreciated. We also want to thank Mr. William Taylor from Drexel University for his significant support of this project.

Conflict of Interest

There are no conflicts of interest.

References

1.
EIA, U.S.
,
2021
, “
How Much Energy is Consumed in Residential and Commercial Buildings in the United States
,” https://www.eia.gov/tools/faqs/faq.php?id=86&t=1, Accessed January 2021.
2.
Katipamula
,
S.
, and
Brambley
,
M. R.
,
2005
, “
Methods for Fault Detection, Diagnostics, and Prognostics for Building Systems—A Review, Part I
,”
HVAC&R Res.
,
11
(
1
), pp.
3
25
.
3.
Roth
,
K. W.
,
Westphalen
,
D.
,
Feng
,
M. Y.
,
Llana
,
P.
, and
Quartararo
,
L.
,
2005
, “
Energy Impact of Commercial Building Controls and Performance Diagnostics: Market Characterization, Energy Impact of Building Faults and Energy Savings Potential
,” Final Report TAIX LLC for the U.S., Department of Energy.
4.
Brambley
,
M. R.
,
Haves
,
P.
,
McDonald
,
S. C.
,
Torcellini
,
P.
,
Hansen
,
D. G.
,
Holmberg
,
D.
, and
Roth
,
K.
,
2005
, “Advanced Sensors and Controls for Building Applications: Market Assessment and Potential R&D Pathways,”
Pacific Northwest National Laboratory
,
Richland, WA
.
5.
Yu
,
Y.
,
Woradechjumroen
,
D.
, and
Yu
,
D.
,
2014
, “
A Review of Fault Detection and Diagnosis Methodologies on air-Handling Units
,”
Energy Build.
,
82
, pp.
550
562
.
6.
Motegi
,
N.
,
Piette
,
M. A.
,
Kinney
,
S.
, and
Herter
,
K.
,
2003
, “
Web-Based Energy Information Systems for Energy Management and Demand Response in Commercial Buildings
,” Lawrence Berkeley National Laboratory, https://escholarship.org/content/qt5z796433/qt5z796433.pdf, Accessed January 2021.
7.
Choi
,
K.
,
Namburu
,
S. M.
,
Azam
,
M. S.
,
Luo
,
J.
,
Pattipati
,
K. R.
, and
Patterson-Hine
,
A.
,
2005
, “
Fault Diagnosis in HVAC Chillers
,”
IEEE Instrum. Meas. Mag.
,
8
(
3
), pp.
24
32
.
8.
Han
,
H.
,
Gu
,
B.
,
Hong
,
Y.
, and
Kang
,
J.
,
2011
, “
Automated FDD of Multiple-Simultaneous Faults (MSF) and the Application to Building Chillers
,”
Energy Build.
,
43
(
9
), pp.
2524
2532
.
9.
Zhao
,
Y.
,
Xiao
,
F.
,
Wen
,
J.
,
Lu
,
Y.
, and
Wang
,
S.
,
2014
, “
A Robust Pattern Recognition-Based Fault Detection and Diagnosis (FDD) Method for Chillers
,”
HVAC&R Res.
,
20
(
7
), pp.
798
809
.
10.
Li
,
S.
, and
Wen
,
J.
,
2014
, “
Application of Pattern Matching Method for Detecting Faults in Air Handling Unit System
,”
Autom. Constr.
,
43
, pp.
49
58
.
11.
Wang
,
S.
, and
Xiao
,
F.
,
2004
, “
Detection and Diagnosis of AHU Sensor Faults Using Principal Component Analysis Method
,”
Energy Convers. Manage.
,
45
(
17
), pp.
2667
2686
.
12.
Yan
,
R.
,
Ma
,
Z.
,
Kokogiannakis
,
G.
, and
Zhao
,
Y.
,
2016
, “
A Sensor Fault Detection Strategy for air Handling Units Using Cluster Analysis
,”
Autom. Constr.
,
70
, pp.
77
88
.
13.
Xiao
,
F.
,
Zhao
,
Y.
,
Wen
,
J.
, and
Wang
,
S.
,
2014
, “
Bayesian Network Based FDD Strategy for Variable Air Volume Terminals
,”
Autom. Constr.
,
41
, pp.
106
118
.
14.
Liu
,
Z.
,
Liu
,
Y.
,
Zhang
,
D.
,
Cai
,
B.
, and
Zheng
,
C.
,
2015
, “
Fault Diagnosis for a Solar Assisted Heat Pump System Under Incomplete Data and Expert Knowledge
,”
Energy
,
87
, pp.
41
48
.
15.
Katipamula
,
S.
, and
Brambley
,
M. R.
,
2005
, “
Methods for Fault Detection, Diagnostics, and Prognostics for Building Systems—A Review, Part II
,”
HVAC&R Res.
,
11
(
2
), pp.
169
187
.
16.
Wiggins
,
M.
, and
Brodrick
,
J.
,
2012
, “
HVAC Fault Detection
,”
ASHRAE J.
,
54
, pp.
78
80
.
17.
Yan
,
Y.
,
Luh
,
P. B.
, and
Pattipati
,
K. R.
,
2017
, “
Fault Diagnosis of HVAC Air-Handling Systems Considering Fault Propagation Impacts Among Components
,”
IEEE Trans. Autom. Sci. Eng.
,
14
(
2
), pp.
705
717
.
18.
Wu
,
S.
, and
Sun
,
J. Q.
,
2011
, “
A Top-Down Strategy With Temporal and Spatial Partition for Fault Detection and Diagnosis of Building HVAC Systems
,”
Energy Build.
,
43
(
9
), pp.
2134
2139
.
19.
Bynum
,
J. D.
,
Claridge
,
D. E.
, and
Curtin
,
J. M.
,
2012
, “
Development and Testing of an Automated Building Commissioning Analysis Tool (ABCAT)
,”
Energy Build.
,
55
, pp.
607
617
.
20.
O'Neill
,
Z.
,
Pang
,
X.
,
Shashanka
,
M.
,
Haves
,
P.
, and
Bailey
,
T.
,
2014
, “
Model-Based Real-Time Whole Building Energy Performance Monitoring and Diagnostics
,”
J. Building Perform. Simul.
,
7
(
2
), pp.
83
99
.
21.
Sun
,
B.
,
Luh
,
P. B.
,
Jia
,
Q.-S.
,
O'Neill
,
Z.
, and
Song
,
F.
,
2014
, “
Building Energy Doctors: An SPC and Kalman Filter-Based Method for System-Level Fault Detection in HVAC Systems
,”
IEEE Trans. Autom. Sci. Eng.
,
11
(
1
), pp.
215
229
.
22.
Georgescu
,
M.
,
Loire
,
S.
,
Kasper
,
D.
, and
Mezic
,
I.
,
2017
, “
Whole-Building Fault Detection: A Scalable Approach Using Spectral Methods
,” arXiv preprint arXiv:1703.07048.
23.
Magoulès
,
F.
,
Zhao
,
H.-X.
, and
Elizondo
,
D.
,
2013
, “
Development of an RDP Neural Network for Building Energy Consumption Fault Detection and Diagnosis
,”
Energy Build.
,
62
, pp.
133
138
.
24.
Qin
,
S. J.
,
2009
, “
Data-Driven Fault Detection and Diagnosis for Complex Industrial Processes
,”
IFAC Proc. Vol.
,
42
(
8
), pp.
1115
1125
.
25.
Liu
,
J.
,
Li
,
G.
,
Chen
,
H.
,
Wang
,
J.
,
Guo
,
Y.
, and
Li
,
J.
,
2017
, “
A Robust Online Refrigerant Charge Fault Diagnosis Strategy for VRF Systems Based on Virtual Sensor Technique and PCA-EWMA Method
,”
Appl. Therm. Eng.
,
119
, pp.
233
243
.
26.
Li
,
G.
, and
Hu
,
Y.
,
2019
, “
An Enhanced PCA-Based Chiller Sensor Fault Detection Method Using Ensemble Empirical Mode Decomposition Based Denoising
,”
Energy Build.
,
183
, pp.
311
324
.
27.
Hassanpour
,
H.
,
Mhaskar
,
P.
,
House
,
J. M.
, and
Salsbury
,
T. I.
,
2020
, “
A Hybrid Modeling Approach Integrating First-Principles Knowledge With Statistical Methods for Fault Detection in HVAC Systems
,”
Comput. Chem. Eng.
,
142
, p.
107022
.
28.
Du
,
Z.
,
Jin
,
X.
, and
Yang
,
X.
,
2009
, “
A Robot Fault Diagnostic Tool for Flow Rate Sensors in Air Dampers and VAV Terminals
,”
Energy Build.
,
41
(
3
), pp.
279
286
.
29.
Hu
,
Y.
,
Chen
,
H.
,
Xie
,
J.
,
Yang
,
X.
, and
Zhou
,
C.
,
2012
, “
Chiller Sensor Fault Detection Using a Self-Adaptive Principal Component Analysis Method
,”
Energy Build.
,
54
, pp.
252
258
.
30.
Li
,
S.
, and
Wen
,
J.
,
2014
, “
A Model-Based Fault Detection and Diagnostic Methodology Based on PCA Method and Wavelet Transform
,”
Energy Build.
,
68
, pp.
63
71
.
31.
Li
,
G.
,
Hu
,
Y.
,
Chen
,
H.
,
Shen
,
L.
,
Li
,
H.
,
Hu
,
M.
,
Liu
,
J.
, and
Sun
,
K.
,
2016
, “
An Improved Fault Detection Method for Incipient Centrifugal Chiller Faults Using the PCA-R-SVDD Algorithm
,”
Energy Build.
,
116
, pp.
104
113
.
32.
Chebel-Morello
,
B.
,
Malinowski
,
S.
, and
Senoussi
,
H.
,
2016
, “
Feature Selection for Fault Detection Systems: Application to the Tennessee Eastman Process
,”
Appl. Intell.
,
44
(
1
), pp.
111
122
.
33.
Ghosh
,
K.
,
Ramteke
,
M.
, and
Srinivasan
,
R.
,
2014
, “
Optimal Variable Selection for Effective Statistical Process Monitoring
,”
Comput. Chem. Eng.
,
60
, pp.
260
276
.
34.
Kohavi
,
R.
, and
John
,
G. H.
,
1997
, “
Wrappers for Feature Subset Selection
,”
Artif. Intell.
,
97
(
1–2
), pp.
273
324
.
35.
Leardi
,
R.
, and
Gonzalez
,
A. L.
,
1998
, “
Genetic Algorithms Applied to Feature Selection in PLS Regression: How and When to Use Them
,”
Chemom. Intell. Lab. Syst.
,
41
(
2
), pp.
195
207
.
36.
Barker
,
M.
, and
Rayens
,
W.
,
2003
, “
Partial Least Squares for Discrimination
,”
J. Chemom.
,
17
(
3
), pp.
166
173
.
37.
Abdi
,
H.
,
2010
, “
Partial Least Squares Regression and Projection on Latent Structure Regression (PLS Regression)
,”
Wiley Interdiscip. Rev. Comput. Stat.
,
2
(
1
), pp.
97
106
.
38.
Wold
,
S.
,
Sjöström
,
M.
, and
Eriksson
,
L.
,
2001
, “
PLS-Regression: A Basic Tool of Chemometrics
,”
Chemom. Intell. Lab. Syst.
,
58
(
2
), pp.
109
130
.
39.
Fujiwara
,
K.
,
Sawada
,
H.
, and
Kano
,
M.
,
2012
, “
Input Variable Selection for PLS Modeling Using Nearest Correlation Spectral Clustering
,”
Chemom. Intell. Lab. Syst.
,
118
, pp.
109
119
.
40.
Sastry
,
K.
,
Goldberg
,
D. E.
, and
Kendall
,
G.
,
2005
, “Genetic Algorithms,”
Search Methodologies
,
E. K.
Burke
and
G.
Kendall
, eds.,
Springer
,
Berlin, Germany
, pp.
97
125
.
41.
Mitchell
,
M.
,
1998
,
An Introduction to Genetic Algorithms
,
MIT Press
,
Cambridge, MA
.
42.
Huang
,
C.-L.
, and
Wang
,
C.-J.
,
2006
, “
A GA-Based Feature Selection and Parameters Optimization for Support Vector Machines
,”
Expert Syst. Appl.
,
31
(
2
), pp.
231
240
.
43.
Murata
,
T.
,
Ishibuchi
,
H.
, and
Tanaka
,
H.
,
1996
, “
Multi-Objective Genetic Algorithm and Its Applications to Flow Shop Scheduling
,”
Comput. Ind. Eng.
,
30
(
4
), pp.
957
968
.
44.
Jain
,
A.
, and
Zongker
,
D.
,
1997
, “
Feature Selection: Evaluation, Application, and Small Sample Performance
,”
IEEE Trans. Pattern Anal. Mach. Intell.
,
19
(
2
), pp.
153
158
.
45.
Siedlecki
,
W.
, and
Sklansky
,
J.
,
1989
, “
A Note on Genetic Algorithms for Large-Scale Feature Selection
,”
Pattern Recognit. Lett.
,
10
(
5
), pp.
335
347
.
46.
Bergmeir
,
C.
, and
Benítez
,
J. M.
,
2012
, “
On the Use of Cross-Validation for Time Series Predictor Evaluation
,”
Inf. Sci.
,
191
, pp.
192
213
.
47.
Wise
,
B. M.
,
Gallagher
,
N.
,
Bro
,
R.
,
Shaver
,
J.
,
Windig
,
W.
, and
Koch
,
R. S.
,
2007
, “
PLS Toolbox 4.0. Eigenvector Research Incorporated
,” Wenatchee, WA, http://mitr.p.lodz.pl/raman/jsurmacki/pliki/zajecia/LMDiT/cw4i5/LMDiT_PLS_Manual_4.pdf, Accessed January 2021.
48.
Hasegawa
,
K.
,
Miyashita
,
Y.
, and
Funatsu
,
K.
,
1997
, “
GA Strategy for Variable Selection in QSAR Studies: GA-Based PLS Analysis of Calcium Channel Antagonists⊥
,”
J. Chem. Inf. Comput. Sci.
,
37
(
2
), pp.
306
310
.
49.
Mehmood
,
T.
,
Liland
,
K. H.
,
Snipen
,
L.
, and
Sæbø
,
S.
,
2012
, “
A Review of Variable Selection Methods in Partial Least Squares Regression
,”
Chemom. Intell. Lab. Syst.
,
118
, pp.
62
69
.
50.
Sakamoto
,
Y.
,
Ishiguro
,
M.
, and
Kitagawa
,
G.
,
1986
,
Akaike Information Criterion Statistics
,
D. Reidel
,
Dordrecht, The Netherlands
.
51.
Weakliem
,
D. L.
,
1999
, “
A Critique of the Bayesian Information Criterion for Model Selection
,”
Sociol. Methods Res.
,
27
(
3
), pp.
359
397
.
52.
Guyon
,
I.
, and
Elisseeff
,
A.
,
2003
, “
An Introduction to Variable and Feature Selection
,”
J. Mach. Learn. Res.
,
3
, pp.
1157
1182
.
53.
Ratanamahatana
,
C.
,
Keogh
,
E.
,
Bagnall
,
A. J.
, and
Lonardi
,
S.
,
2005
, “
A Novel Bit Level Time Series Representation With Implication of Similarity Search and Clustering
,”
Pacific-Asia Conference on Knowledge Discovery and Data Mining
,
Hanoi, Vietnam
,
May 18–20
,
Springer
, pp.
771
777
.
54.
Keogh
,
E.
,
Chakrabarti
,
K.
,
Pazzani
,
M.
, and
Mehrotra
,
S.
,
2001
, “
Dimensionality Reduction for Fast Similarity Search in Large Time Series Databases
,”
Knowl. Inf. Syst.
,
3
(
3
), pp.
263
286
.
55.
Lin
,
J.
,
Keogh
,
E.
,
Wei
,
L.
, and
Lonardi
,
S.
,
2007
, “
Experiencing SAX: A Novel Symbolic Representation of Time Series
,”
Data Min. Knowl. Disc.
,
15
(
2
), pp.
107
144
.
56.
Lin
,
J.
,
Keogh
,
E.
,
Lonardi
,
S.
, and
Chiu
,
B.
,
2003
, “
A Symbolic Representation of Time Series, With Implications for Streaming Algorithms
,”
Proceedings of the 8th ACM SIGMOD Workshop on Research Issues in Data Mining and Knowledge Discovery
,
San Diego, CA
,
June 13
,
ACM
, pp.
2
11
.
57.
Zhu
,
M.
, and
Ghodsi
,
A.
,
2006
, “
Automatic Dimensionality Selection From the Scree Plot via the Use of Profile Likelihood
,”
Comput. Stat. Data Anal.
,
51
(
2
), pp.
918
930
.