Detection and elimination of insignificant interacting subsystems in MIMO closed-loop systems using the least mean square-based partial correlation algorithm

Closed-loop identification of multi-input multi-output (MIMO) systems in large-scale plants has significant difficulties due to subsystem interactions. This complexity is attributed to several input‒output variables, interactions such as recycling to improve or save material and energy, and disturbances such as heating or cooling within the plant. One of the fundamental problems in closed-loop identification is the input perturbation of the interacting subsystems to capture the dynamics of the system for producing an informative dataset and consequently obtaining an accurate model. However, perturbing all the interacting subsystems in the plant increases the applied excitation signals, which makes the identification a nontrivial task. Thus, a precise and quantitative procedure to evaluate the significance and contribution of such interacting subsystems before applying these excitation signals is required to simplify the identification task. Conventional partial correlation analysis is one of the implemented techniques to measure the significance of these interacting subsystems. However, this technique is based on least square estimation. Thus, incorrect estimation of the model errors is produced due to the correlations amongst the process inputs and unmeasured disturbances. Accordingly, this paper describes the implementation of a developed least mean square-based partial correlation algorithm for detecting and eliminating insignificant interacting subsystems of MIMO closed-loop systems. The developed algorithm can discriminate the interacting subsystems that substantially influence the plant interaction from those that do not by minimizing the model regression errors produced due to the process input correlation, unmeasured disturbances, and colored noise. The effectiveness of the proposed method is demonstrated through a case study.


Introduction
Several industrial systems and plants are multivariable by nature.The complex nature of identifying such processes arises from the embedded interactions amongst the process inputs and outputs.Thus, each control loop is necessarily affected by other loops, and thus, the control of such a process becomes challenging [1,2].Moreover, such processes generally suffer from perfectly correlated input signals, low signal-to-noise ratios (SNRs) [3], correlated noise signals, and colored noises.
Simultaneous perturbation of all external reference signals is adopted to enrich the identifiability of a model structure and deliver an informative dataset in multi-input multi-output (MIMO) systems [4].However, perturbing only one or a few input signals leads to a simple implementation for the identification task [5].Moreover, introducing such perturbing signals is usually unattractive or too costly because such signals typically interrupt plant operation and consequently disturb product quality [6].
Such complex simultaneous perturbations are often undesired, although they can excite any system.This situation can be attributed to random or large excitations that lead to excessive process variability and make the operation personnel anxious [7].Therefore, defining appropriate settings that permit the design of the minimal signal properties that are essential for the reference signal and obtaining an accurate system model while reducing the effect of the perturbations are desirable [8].
The results in [5,9,10] revealed that signal perturbation is optional to accomplish the identification of a system.Noise sources, as an alternative, can deliver an informative dataset.In such cases, the selected controller has adequate complexity compared with the selected model structure in MIMO systems under closed-loop control [11].However, a lengthy dataset must be acquired to fulfil the given accuracy levels.In [5], the results of not exciting all reference signals were validated by applying variance analysis.
In [12], the influence of reducing the sampling rate on discrete, routine operating closed-loop system identification was investigated.The data were collected without external excitation and were exclusively affected by natural disturbances.Recently, typical results for identifiability were determined by applying closed-loop data with or without modifications in the reference input signal [8].The study is an extension of the work presented in [12].The sampling time and time delay are dominant factors that define the difficulty level of the applied reference signal.
All the abovementioned works show the necessity for simplifying the identification task by minimizing the reference signal perturbations.However, the possible reduction in input signal perturbation can be further investigated.The input variables of the interaction transfer functions are the most suitable candidates given that some of them hold very minimal dynamic information about large-scale plants.Thus, they can be left out from being perturbed according to correlation analysis and then let off from the identification task [13].
In reality, exciting all the input signals, irrespective of their contribution to interaction, is impractical for the MIMO closed-loop system case because of process restrictions or economic reasons.Thus, the interacting subsystems that contribute to interaction should be eliminated in a primary stage to ensure that minimal input signals are perturbed.Subsequently, the MIMO system can be identified by using any identification method, such as the prediction error method (PEM).
The determination of the null or less contributing input-output pairs in the multivariable system was investigated in many studies by determining the controllability based on the minimum singular value decomposition analysis in defined frequencies [14][15][16][17].Although these methods have successful applications in open-loop systems, they generally provide misleading results when applied to detect and eliminate no-model pairs in MIMO closed-loop systems [18].This issue is attributed to the feedback mechanism and the controller action [19].Controller action commonly results in a decoupled situation that hides information regarding model existence for the input-output pair [20].
Recently, several studies have been conducted to overcome the controller action and inherent feedback in closed-loop systems using correlation-based methods.In [18,19,21], the determination of input-output combinations in MIMO closed-loop systems with no-model pairs is presented based on a correlation analysis method.In [21], the low estimated values of the polynomial coefficients of the identified auto-regressive with eXogenous input (ARX) model are the key elements to deciding the no-model pairs.However, the associated limitation with this method is the requirement of prior knowledge of the controller law of each subsystem, which is due to that the control law is part of the correlation analysis.
In [19,21], Pearson's correlation analysis between the set point and the controller output signals is implemented as part of the algorithm implementation.However, the implemented correlation analysis based on Pearson's correlation certainly provides misleading results in the MIMO closed-loop implementation due to the inherent and hidden input-output variables correlation due to the feedback nature.Thus, implementing the partial correlation analysis is effective compared with the regular Pearson's correlation due to the possibility of the input-output variables being correlated [13].
The partial correlation method can be applied to remove the effect of other correlative variables by keeping them untouched in the investigation.Thus, the correlation between the two variables of interest is known [22].The partial correlation technique is trusted if one ensures that the regression error terms for conducting the analysis are unaffected by variable correlations or unmeasured disturbances, which is not the case in multivariable industrial practices.Another difficulty arises in the conventional partial correlation because of the perfect correlation between the process input signals from the neighbor subsystems.The aforementioned problems make the assessment of the dependency between these input signals using the conventional partial correlation analysis misleading due to the use of the least square (LS) method in calculating the model regression errors.
In [13], the differentiation between the interacting subsystems was achieved using conventional partial correlation analysis where the LS method was adopted.Complications, such as bias estimation of the model regression error terms because of the unmeasured disturbances and correlations that occur amongst the process input signals from the other subsystems in the plant, constrain the success of the partial correlation-based LS method.
The least mean square (LMS) method is a popular adaptive algorithm that was proposed in [23].The algorithm was applied iteratively to overcome the local minima and slow learning problems associated with backpropagation in [24] and for other nonlinear applications [25,26].The LMS algorithm utilizes the unbiased instantaneous gradient that is determined based on the most recent information.Thus, minimizing the instantaneous squared error criteria recursively using an LMS algorithm can be effective in partially removing the effects of correlations between the disturbances and the manipulated variables and results in unbiased and consistent parameter estimates [11].In addition, this instantaneous parameter estimate avoids the necessity for the prefiltering process of the experimental data.Thus, it evades the possible loss of some of the system dynamics due to the improper selection of frequency ranges.
In this study, the LMS-based partial correlation algorithm is developed to avoid the restriction of the conventional partial correlation.The LMS method combines the simplicity of LS and the effectiveness of the PEM.
The algorithm guarantees accurate results, while the conventional partial correlation probably delivers misleading results in particular cases.Thus, in this study, determining which subsystems are unimportant and thus left out of the identification process at a prior stage makes the task of selecting the input signals to be perturbed straightforward.

Methods
A multivariable process under decentralized control is considered, as shown in Fig. 1.
For the closed-loop setting in Fig. 1, the case of three decentralized loops is considered, and the aim is to assess the importance of the individual interacting dynamics and their contribution to the subsystem interaction under closed-loop conditions.In this case, the interacting subsystems (in the red rectangle) affecting loop 2 are examined for example.Thus, the interaction from neighbor subsystems (g 21 and g 23 in this case) has to be investigated.The question now is whether both of these interaction transfer functions directly affect the closed-loop identification in loop 2 or whether one of these interaction transfer functions depends on the other due to some correlation.
Conventional partial correlation analysis can be applied to examine the dependency relationship amongst u 1 , u 2 , and u 3 for answering the abovementioned question.Firstly, the computed error terms from the solution of the linear regression problem between these inputs represent the independent component for each input signal.Secondly, another linear regression problem is to be solved between (y 2 ) and the inputs (u 1 , u 2 , and u 3 ) to compute the output error.The computed errors in the first and second stages are lastly used to compute the partial correlation coefficients following Eq.( 1).If the test shows any dependency, then the specific transfer function has less influence on the interaction and could be eliminated.Consequently, the identification task is reduced, and the identification precision is enhanced by eliminating the interacting subsystems to be dithered.
The partial correlation can be represented as a correlation between (u 1 , y 1 ): where n is the number of data points and σ is the variance of the analogous subscript.

Development of the LMS algorithm for MIMO systems
The linear regression form for the information vector and the parameter matrix is regarded as follows: The parameter matrix is defined as follows: where n = q × n a + p × n b and the corresponding information vector is as follows: where are the process input and output vectors, respectively.Recursive LMS algorithms are commonly used online to update the parameter estimate vectors as time t increases as illustrated in Fig. 2. The algorithm applies the mean square error (MSE) criterion to minimize the cost function while estimating the parameter vector.
For the sake of compactness, ε(t, θ) is denoted as ε hereinafter.Then, the MSE crite- rion function in Eq. ( 5) above can be defined as follows: (1) The algorithm uses the gradient descent technique to attain the optimal Wiener solution θ * (t) by minimizing an instantaneous squared error that is quadratic in the param- eter vector.Thus, it guarantees the existence of a single minimum in the error surface.Accordingly, the algorithm converges to this single minimum.
Minimizing J [θ(t)] in Eq. ( 6) delivers the LMS estimate of θ(t) , with the assumption that ϕ(t) is persistently excited using a perturbation signal.The recursive equation to compute the parameter update θ(t − 1) is given by [27,28].
where µ > 0 is the step size parameter that controls the convergence rate and stability of the algorithm.The new estimate of θ(t + 1) is determined by the gradient of J θ(t) according to θ(t): Equation ( 8) will be used to drive the LMS updating equation, with the assumption that the noise vector v(t) is an independent identically distributed (i.i.d.) sequence and the information vectors ϕ(t) are independent over time.The first assumption is wide- spread, but the second assumption is far from being realistic and appears to be difficult to hold in most of the applications [29] and not even true for the input-output pairs in ϕ(t) for t = 1 to N, where N is the number of data points [30].
The reason is that a large part of the data are shared between two successive information vectors, and therefore, they are strongly correlated.Accordingly, the convergence analysis for the case where the assumption is invalid, and the correlated observations used in ϕ(t) were investigated in several works.In [29], the convergence analysis was conducted under assumptions close to practical applications, where a (6) Fig. 2 LMS setup for MIMO system strong correlation in ϕ(t) was considered.The authors showed that the mean square difference between the optimal and true parameter estimate vectors leans towards zero with the decrease in the step size parameter µ.This result is similar to that in [24,31], where the analysis shows that the LMS solution converges to a stable range when the step size approaches zero.In [32], the convergence under the common assumption of correlated observations was also investigated.The author proved that when µ approaches zero, the convergence of an algorithm is nearly sure.
Thus, the cost function J θ(t) is minimized based on the instantaneous estimate of ε computed by the following: By replacing θ(t) in Eq. (11) with its iterative parameter estimates, θ k (t) , the esti- mate of ε can be iteratively computed at iteration k by the following: In the iterative LMS algorithm, the parameter estimate equation updates are iteratively computed as follows: The gradient of the iterative LMS cost function will be as follows: Equation ( 15) is the MIMO counterpart of the LMS for the single output case in [31,33].This equation shows that the LMS algorithm can improve the parameter estimate accuracy due to the implementation of the iterative procedure.( 9)

Elimination procedure of the interacting subsystem using the LMS-based partial correlation algorithm
Step 1: The effect of all other variables in U 1 on u i is eliminated.

The linear regression is written as follows:
where U is an n × p matrix of causal variables, θ u 1 is a p × 1 vector of the parameters to be estimated, and u i and ei are the p × 1 vectors of the effect variable and regression errors, respectively.
Given that U contains all other subsystem input variables u j (j ≠ i), the LMS expres- sion to minimize the cost function is obtained by applying Eq. ( 15) for the U matrix: The regression errors for the abovementioned equation are estimated as follows: e ui is a vector of the independent components of u i that neglects the effects of all other variables.
Step 2: The impact of U on y i is estimated by a regression model: where y i is the output of the main subsystem from which θ y i can be estimated as follows: The regression errors for y i can then be estimated as follows: Step 3: The regression error terms are assumed to involve those elements of u i and y i that are independent of U. Therefore, the partial correlation can be represented as a regular correlation between (u i ,y i ): where N is the number of data points and σ is the standard deviation of the corresponding subscript.(16) Given that the tests will be conducted in a dynamic sense, the value of N can be selected to be adequately large based on the process knowledge to consider the presence of the time delay in the signals under analysis.
For conventional partial correlation analysis, if the true regression model is given by the LS method as follows: Then, the parameter estimate errors can be computed as follows: If N approaches infinity: In LS estimates to maintain θ − θ 0 LS = 0 , one has to keep E U T e i = 0 , and this condition is only satisfied if e i is white noise.Therefore, the LS usually has nonzero asymptotic bias and is not consistent with the case of colored noise, as the case in real MIMO systems operates under closed-loop control.However, estimating the regression error on the basis of LMS in this case is certainly more accurate.
The key difference between the conventional partial correlation method and the LMS-based partial correlation algorithm is the way to calculate the regression error vectors e ui and e yi that contain the cancellation of the impact of all other variables on u i and y i .In the conventional partial correlation method, a simple linear regression problem is firstly solved to calculate the parameter estimates θ u ik and θ y ik .Then, the regression error vectors e ui and e yi are obtained by solving two different regression models.Lastly, Eq. ( 22) is applied to calculate the partial correlation coefficients.This case is different from that of the proposed LMS-based partial correlation algorithm in this study, where the LMS algorithm is used to accurately determine the regression errors.The accurate calculation of the regression errors will obviously lead to more accurate partial correlation coefficients given that Eq. ( 22) heavily depends on these regression errors.

LMS-based partial correlation algorithm
i. Collect input-output data from the closed-loop experiment.ii.Solve the linear regression problem between u i and the input matrix U. iii.Solve the linear regression problem between y i and the input matrix U. iv.Compare θ u ik (N ) and θ y ik (N ) with θ u ik−1 (N ) and θ y ik−1 (N ) , respectively: if they are appropriately close to the predetermined small ε, then terminate the procedure and obtain the iterative estimates at k; otherwise, increase k by 1. v. Determine the regression errors for the two regressions e ui and e yi above.vi.Calculate the partial correlation coefficients using Eq.(22).vii.Apply the t-test for significance.(23)

Case study description
This case study considers two different simulation scenarios of white and colored noises affecting the plants.The case study is applied to demonstrate the power of the developed LMS-based partial correlation algorithm for interacting subsystems elimination in strongly interacting MIMO closed-loop systems compared with that of the conventional partial correlation method.

Copolymerization process
The transfer function model in Eq. ( 26) of a copolymerization process of methyl methacrylate and vinyl acetate in a continuous stirred tank reactor (CSTR) reported in [34] has been considered in this study.
The best pairing is determined by the RGA analysis.The system was divided into three decentralized control loops.In the first decentralized loop, the pairing was y 1 , y 2 with u 2 , u 3 .The suggested pairing for the second loop was y 3 with u 4 , and for the third loop, it was y 4 with u 5 , while u 1 was kept constant.Proportional controllers k c1 , k c2 , k c3 , and k c4 with the values 0.1, − 0.3, 0.5, and − 0.5 were used respectively to stabilize the control system.To compare the conventional partial correlation against the LMSbased partial correlation, the process model in Eq. (26)

Results and discussion
Two simulation cases are illustrated to demonstrate the cases in which the LMS-based partial correlation is superior compared to the conventional partial correlation.In the first simulation case (referred to as case A), the process is affected by white noise, assuming smooth plant operation.While in the second simulation case (referred to as case B), a more complex case that represents the real disturbances in large-scale plants is considered; hence, a colored noise was added.The colored noise was produced by filtering the ( 26) added white noise sequences to the process outputs and simulated by the following firstorder filter described in Eq. (27).
where w 1 (t) to w 4 (t) are assumed to be produced by filtering the white noise sequences v 1 (t) to v 4 (t), respectively.The white noise has zero mean and variances σ 2 equal to 0.086, 0.066, 0.043, and 0.072 for v 1 (t) to v 4 (t), respectively.The SNR values for the process outputs in the two simulation cases are reported in Table 1.In this analysis, 3500 data points were generated with a sampling interval of 1 min to conduct the partial correlation analysis for the conventional and the proposed algorithm.
In case A, white noise is introduced to the process.The effect of the white noises on the process outputs in case A is shown in Fig. 3, whereas Fig. 4 shows the sequential setpoint step changes applied to the process outputs of cases A and B.
While in case B, a colored noise was introduced to reflect a similar situation in largescale plant affected by several disturbance sources.Figure 5 demonstrates the characteristic of the added colored noise to the process outputs.The figure shows the power ( 27)  To retain the real case of plants operated under closed loop, the sequential step changes in Fig. 4 were applied to all process outputs in the two simulations cases A and B in Figs. 3 and 6, respectively.These changes were applied to each output separately, starting from y 1 and moving to the other outputs after a pre-specified period.The sequential step change might be necessary in this case to avoid any possible correlation in the reference signals and subsequently adding more bias to the regression error terms, which negatively affect the accuracy of the estimated partial correlation coefficients.
In this paper, the effect of different characteristics of white and colored noises on the concerned plants has been investigated.Hence, the PSD is a crucial tool adopted in signal frequency analysis to define the signal power distribution over the frequency and is generally applied in the identification theory to distinguish between different noises characteristics.
The white noise is regarded as an ideal stationary random signal and has a constant power spectral density across the entire frequency spectrum.For the white noise case, it is known that a flat profile is generally obtained since no signal frequency contributes extensively to comparison with another one.Hence, the correlation or autocovariance between the samples is zero through all the sampling instances.
In the least square estimation method, the best parameter estimates can be found conditioned on an existing noisy dataset.Consequently, when random noise is in the measurements, as shown in Fig. 3, this conditioning removes the randomness from the concerned problem.It considers independent and identically distributed (i.i.d) noise.Therefore, the performance of the LS method is satisfactory in such cases, and subsequently, a reliable partial correlation coefficient can be produced by the conventional method.However, when the plant is affected by low SNRs, the performance of the LS method is certainly degraded [2,3,35], and therefore, the conventional partial correlation coefficients will be misleading.
The system in this case study is strongly interacting, and this can be observed from the set point applied to 2 and y 4 , as shown in Fig. 4.Although the set point for this particular output is applied with positive magnitude, however, the output response is in the opposite direction, as illustrated in Fig. 3 (downward).This is attributed to the fact that the second decentralized loop in this plant has the lowest dynamics according to the low-order transfer functions in this loop.This makes the contribution of other interacting subsystems more dominant compared to the dynamics of the loop itself.On the other hand, loop 4 has no interaction effect from other subsystems; therefore, the corresponding output y 4 has immediately responded to the applied setpoint change at 2750 (upward) since the process does not incorporate a process time delay.
In case B, Fig. 5 shows a colored noise without uniform power spectral density across the entire frequency spectrum, which reflects a similar situation in large-scale plants affected with several disturbance sources.Therefore, the correlation or the autocovariance between the samples is non-zero through all the sampling instances.
Adopting the LMS algorithm to the outputs affected with colored noise as illustrated in Fig. 6 overcomes the issue of the correlation between the colored noise samples as it minimizes an instantaneous squared error criterion and does not rely on cross or autocorrelations between the samples.Moreover, the LMS algorithm assumes that the measures in the cost function are affected with random noise.Thus, a noise-generating process is applied through the parameter estimation process using an unbiased instantaneous gradient at each instant, and subsequently produces more accurate regression errors, which are the key component in computing the partial correlation coefficients.
This situation differs from the white noise case where an ideal stationary random signal is assumed with a zero correlation between the samples through all the sampling instances affecting the plant.Therefore, in case B, biased and inconsistent estimates can be produced by the LS method and, consequently, incorrect partial correlation coefficients.
On the case study considered in this section, the cause and effect relationship was analyzed using conventional partial correlation and LMS-based partial correlation methods.The analysis was applied to full plant dynamics for the normal operation case in case A and to a more complex scenario in case B due to the addition of the colored noise, and the results are illustrated in Figs.7 and 8, respectively.
The results in Figs.7 and 8 show significant contribution of all the interacting subsystems, and none of them can be isolated prior to the identification stage; thus, an assumption is made to intentionally remove one of these interacting subsystems from the process transfer function model to assess the performance of the two partial correlation algorithms when different types of noise are added to replace the intentionally removed interacting subsystem.However, the interaction transfer function (y 3 -u 2 ) is removed (assumed to be zero), and new data are collected while replacing the removed interacting subsystems with white and colored noise in cases A and B, respectively.This assumption can also demonstrate to what extent the LMS-based partial correlation is able to differentiate between the actual correlation between (y 3 -u 2 ) and that between y 3 and the white and colored noises after the interacting subsystem is removed.The corresponding results are presented in Figs. 9 and 10 for cases A and B, respectively.
The significance of the partial correlation coefficients for the full plant analysis in Figs.7 and 8 has been verified by applying the t-test, and Tables 2 and 3 reflect the t-test results.Similarly, Tables 4 and 5 reflect the t-test results of the partial correlation coefficients in Figs. 9 and 10 for the plant with isolated interacting subsystems analysis.
A value of zero was produced by the t-test when the LMS-based partial correlation algorithm for the removed transfer function (y 3 -u 2 ) was applied, as it no longer existed in the two simulation cases, while the t-test shows a value of one for the conventional partial correlation in case B. This indicates that the conventional partial correlation can produce correct results in normal operating conditions while not under colored noise conditions.
In case A, at less disturbed conditions for the full plant case, comparable results were obtained, and small partial correlation coefficients for the interaction transfer functions were reported for both the conventional partial correlation and the LMS-based partial correlation algorithm, as shown in Fig. 7. Nevertheless, under colored noise conditions in case B for the same full plant analysis, both partial correlation methods reported slightly higher partial correlation coefficients as shown in Fig. 8.
For the input-output control pairing (within the dash-dot lines in the figures), small partial correlation coefficients were obtained by the conventional partial correlation noise.In contrast, the LMS-based partial correlation algorithm was able to differentiate between the removed correlation between u 2 and y 3 and between the one that is due to the colored noise as shown in the same figure.
On the other hand, the LMS-based partial correlation algorithm shows less partial correlation coefficient values for most of the interaction transfer functions in both simulation cases compared with the conventional partial correlation, which demonstrated high partial correlation values mainly in case B. This indicates that the LMS-based partial correlation algorithm is able to estimate the regression errors more accurately by eliminating the error due to the correlation between y 3 and the added colored noise and, consequently, resulted in a partial correlation coefficient that only reflects the causality relation.
Figure 11 shows the average regression errors that were calculated by the proposed LMS-based partial correlation algorithm for the process outputs for the full plant analysis.For case A, the algorithm required 3 iterations for error convergence (except y 4 ), while more iterations were required for the error convergence for all the outputs in case B.Moreover, the final regression errors after the last iteration in case B are higher than that in case A. This is attributed to the fact that when the plant is affected with white noises, the algorithm can easily compute the regression errors, while the algorithm is encountering some numerical difficulties when colored noise or low SNR affect the plant.
Finally, Fig. 12 shows higher regression errors for the conventional partial correlation method compared to that of the LMS-based partial correlation algorithm with the isolated interacting subsystem.Usually, the regression errors only represent the unmodeled part that comes with no information about the cause and effect relationship in the system under investigation.However, when the LS method is used, the regression errors may contain an additional component that represents the correlation between the dependent variables and some colored noise dynamics, and thus, the actual regression error values increase in an incorrect way.
Recall Eq. ( 28), the partial correlation coefficients heavily depend on the accurately estimating the regression errors e ui and e yi .Thus, when significant regression errors are misestimated, as in the case when the system is highly correlated or affected by colored noise, the numerator value of Eq. ( 28) becomes much bigger, and consequently, the partial correlation coefficients increase, providing misleading and confounding partial correlation results.However, the main difference between the conventional partial correlation method and the LMS-based partial correlation algorithm is in calculating these regression errors more accurately using the LMS algorithm.

Fig. 3 Fig. 4 Fig. 5
Fig. 3 Process outputs of case A with the addition of white noise

Fig. 6
Fig. 6 Process outputs of case B with the addition of colored noise

Fig. 7 Fig. 8
Fig. 7 Conventional and LMS-based partial correlation analysis for full plant analysis (A)

Fig. 9 Fig. 10
Fig. 9 Conventional and LMS-based partial correlation analysis for plant with isolated interacting subsystem (A)

Fig. 11
Fig. 11 Average regression errors for cases A and B

Table 1
SNR values for the process outputs in the two simulation cases

Table 2 t
-test results of conventional and LMS-based partial correlation for full plant analysis (A) t-

Table 3 t
-test results of conventional and LMS-based partial correlation for full plant analysis (B)