We consider a n × p data matrix X in which the n rows are the observation vectors , . Letting denote the mean vector of dimension p, the covariance matrix is involved in a wide range of statistical methods including principal component analysis or multiple regression for example. Since S is known to be highly prone to influential observations, sensitivity aspects for principal component analysis have been discussed in several papers including Critchley (1985), Jolliffe (2002), Pack et al. (1988), Prendergast (2008), Prendergast and Li Wai Suen (2011), Tanaka (1988) among many others. It should be noted that perturbation issues for this method are still the subject of active research as emphasized by the recent work of Masioti et al. (2023) for example. In order to assess the influence of a small subset I of r observations on the eigenstructure of S, a possible approach consists of studying the effect of removing these r observations on the eigenelements of S. If we note a significant modification of the eigenelements, we know that these observations are strongly influential on the results. In this framework, letting denote the covariance matrix obtained without the subset of observations indexed by I, several authors have studied the relationship between the eigenelements of S and those of . More specifically, providing approximations to the eigenelements of allows to detect influential subsets of observations without having to recompute the exact modified eigenvalues and eigenvectors. Hadi and Nyquist (1993), Masioti et al. (2023), and Wang and Nyquist (1991) study the effect of deleting a single observation while Bénasséni (2018), Enguix-González et al. (2005) and Wang and Liski (1993) focus on the general case where I comprises several observations. In these works, approximations to the eigenvalues and eigenvectors are obtained by retaining the first terms in power expansions of these parameters. Independently of these works, Bénasséni (1987) suggests using approximations based on Rayleigh quotients together with inequalities given in Wilkinson (1988). More generally, approximations for the covariance matrix are also the subject of research in a wider framework as emphasized by other works including for example Enguix-González et al. (2015) which considers the moments of the eigenelements or Enguix-González et al. (2012) dealing with the conditional bias of eigenvalues.
The contribution of this work is twofold. First, some new elements of comparison are provided between approximations derived from power expansions and those based on Rayleigh quotients. This comparison is based on some simple theoretical relations together with the numerical study of a data table which is intended to provide some guidance on the approximations to choose in practice. Second, refined inequalities in Chatelin (2012) are introduced in order to evaluate the accuracy of approximations without having to recompute the eigenelements of . It is proved that these inequalities always provide a sharper evaluation than those used in Bénasséni (1987). This is a definite advantage from a computational standpoint.
In this work, the eigenvalues of S are assumed simple and associated to the normalized eigenvectors . In the same way, the eigenvalues of the perturbed matrix are also assumed simple and associated to the eigenvectors .
Approximations Based On Power Expansions
Theoretical Background On Matrix Perturbations
Referring to Bénasséni (2018) or Enguix-González et al. (2005) and letting , we know that, when a subset of observations indexed by I is deleted, the covariance matrix S is transformed to which can be expressed as:
1
We then have a perturbation of the form with , and . Following matrix perturbation theory detailed in Wilkinson (1988) for example or referring to Sibson (1979), we know that, if ϵ is sufficiently small, for each simple eigenvalue λ of S there is an eigenvalue of given by a convergent power series:
2
with a corresponding eigenvector which can also be expressed under a convergent power series:
3
The parameters γ1, γ2, …, γm and ,,…, are derived by equating the coefficients of ϵ, ,…, in the equation . It should also be noted that the perturbation of λk may not necessarily be the kth largest eigenvalue of if the subset of observations indexed by I has initially a strong influence on λk for example. In this case, following Critchley (1985), one can simply assume that the eigenvalues have been reordered in decreasing order and that their corresponding eigenvectors have been relabeled. However, the reader is referred to Masioti et al. (2023) for a more comprehensive discussion on this topic.
Formulae for the Approximations Derived From Power Expansions
For any integer , retaining only terms of order lower or equal to m in ϵ in (2) and (3) provides the approximations and of order m for the eigenvalues and eigenvectors of .
This is the general approach suggested in Bénasséni (2018), Enguix-González et al. (2005) and Wang and Liski (1993). Assuming that is sufficiently small to ensure the convergence of the above power series, Enguix-González et al. (2005) provide the following approximations for :
4
5
6
where , for and .
Details on the derivation of the above expressions are given in Bénasséni (2018). The reader is referred to Enguix-González et al. (2005) for the formula of which is fairly long and therefore omitted in this paper. It should also be pointed out that a comprehensive study of approximations for the unbiased matrix is provided in this last reference. In the remaining of the paper, we define also the approximations of order zero as and for notational convenience.
Finally, when studying the influence of a single observation, note that approximations are simply obtained by taking r = 1, and in the previous developments. The reader more specifically interested by this case will refer to a series of papers including Critchley (1985), Hadi and Nyquist (1993), Masioti et al. (2023) and Wang and Nyquist (1991).
Approximations Based On Rayleigh Quotients
Rayleigh Quotients as Approximations to the Perturbed Eigenvalues
Assuming that weights are given to the observations, Bénasséni (1987) studies the effects of modifying these weights on the eigenvalues and eigenvectors of the covariance matrix. Deleting a small subset of observations indexed by I is therefore a particular case of his approach which consists simply of modifying to zero the corresponding weights. As approximations to for , this author suggests using the Rayleigh quotient for and the initial normalized eigenvector , and the Rayleigh quotient for and the approximation of order one to .
Error Analysis
From a computational standpoint, it is of major interest to evaluate the accuracy of approximations without having to recompute the exact eigenelements of . In order to do this, Bénasséni (1987) suggests using inequalities provided in Wilkinson (1988). Focusing on and its corresponding eigenvector and, from now on, assuming without change of notation that has been normalized, let for where stands for the two norm. Assume that is a nonzero positive constant such that for with . Then the accuracy of as approximation to , and of as approximation to , is analyzed in Bénasséni (1987) using the following inequalities given in (Wilkinson, 1988, pp. 172–176):
7
and if :
8
Using , note that (7), can be written with the cosine between and as:
9
In practice, it is necessary to give the parameter a value in the previous inequalities. When studying the influence of a single observation, it should be noted that (1) can be expressed as
10
so that we have a rank one perturbation. In this case, the parameter is given a value in Bénasséni (1987) using bounds, derived from the Courant-Fischer theorem, for the eigenvalues of the symmetric matrix , , . In the case where several observations are deleted this author suggests a fairly lengthy procedure assuming that these observations are removed one after the other, so that we have a series of rank one perturbations.
New Developments
Some Relations Between the Approximations
In his work, Bénasséni (1987) only considers the approximations of order one for the eigenvalues of and provides no comparison of with approximations based on Rayleigh quotients. However, it is easy to derive the following simple relations. First note that can be written as:
11
using (1). Therefore we see that we have always . Furthermore, a simple comparison of (11) with (5) shows that:
In particular, when focusing on the largest eigenvalue which plays a central role in PCA, this difference is non negative so that we have . In a similar way, when considering the smallest eigenvalue, we get .
We omit the derivation of which is tedious and leads to a formula too complicated to be interpreted. However, it should be noted that this approximation involves terms up to the order 4 in .
Improved Inequalities in Error Analysis
In practice, it is of crucial importance to have bounds as close as possible to the true values of and in Inequalities (8) and (9). This will allow for an evaluation as precise as possible of the real accuracy of the approximations without having to recompute the perturbed analysis. For this purpose, we introduce below refined inequalities in the study of covariance matrices in order to get an improved error analysis
Indeed, Inequalities (7), (8) and (9) introduced in the Subsection Error Analysis can be improved using error analysis developed in Chatelin (2012, pp. 180–184). More precisely, it is easily derived from Corollary 4.6.4 in this reference that:
12
and
13
under the condition:
14
It is obvious that (12) is more accurate than (8). A similar remark holds for (13) wich improves (9). This last point is easily checked by converting (13) into
15
Then letting , and , Inequality (9) becomes so that we have simply to prove that . Developing B, we get . Thus we have that if and only if the polynomial is non-positive. This is the case when a belongs to . Therefore the result follows since from (14).
Furthermore, when dealing with the eigenvector associated to the largest eigenvalue, Chatelin (2012, p. 204) points out that Inequality (13) can also be refined into the following tangent based inequality:
16
since this eigenvalue is assumed to be simple. More precisely, letting α denote the angle between the two vectors and , we have since for all . Thus we obtain a better approximation of α when using the function arctan rather than the function arcsin showing that (16) improves (13).
Improved Value for the Parameter
The sharpness of the bounds in Inequalities (12), (13) and (16) depends on the value of the parameter . The larger is, the sharper are these inequalities. In order to get a suitable value for this parameter, some other results of Chatelin (2012, pp. 180–181) or of Wilkinson (1988, pp. 174–176) turn out to be also of specific interest since they often improve significantly the value of the parameter obtained through the Courant-Fischer theorem in Bénasséni (1987). More precisely, from these references we know that there is at least one eigenvalue of in each of the intervals defined for by which are often referred to as Krylov-Weinstein intervals. When the interval is isolated from the other ones, we know that it contains precisely one eigenvalue. Then, assuming that the Rayleigh quotients satisfy (after having been reordered if necessary), a value for can be easily derived as if , if k = 1 and if . Finally, note that these Krylov-Weinstein intervals proved to be also useful in studying the effects of some small errors in the data table itself as emphasized by the work of Bénasséni (1988, pp. 303–310)
It should be pointed out that for very close eigenvalues, giving a value to this parameter can be a real issue. However, once a value satisfying (14) is obtained, we know from (12) that, for , the eigenvalues lie in the intervals which are sharper than the Krylov-Weinstein intervals as soon as (14) holds. These new intervals can be used to obtain a larger value for the parameter , thus improving (12), (13) and (16). This process could be iterated, but no significative improvment is generally observed.
Finally, it was noted in (10) that we have a rank one perturbation when deleting a single observation. Several inequalities and relations for this restricted rank perturbation are given in Bénasséni (1987), Hadi and Nyquist (1993), and Wang and Nyquist (1991). The reader is also referred to more recent works by Bénasséni (2011), Cheng et al. (2014), and Ipsen and Nadler (2009), who suggest new bounds for the perturbed eigenvalues. Although Krylov-Weinstein intervals generally provide a satisfying value for the parameter , these recent works can also be interesting in the determination of the largest possible constant in order to make Inequalities (12), (13) and (16) sharper.
Numerical Study
The numerical illustration of the results is based on the soil composition data in Kendall (1975) which have already been used in several works including, among others, Bénasséni (2018), Critchley (1985), Enguix-González et al. (2005), Enguix-González et al. (2012), Tanaka (1988), Wang and Liski (1993), or Wang and Nyquist (1991) for sensitivity study of covariance based principal component analysis. The data table consists of 20 observations measured on 4 variables. We have the following four eigenvalues for the corresponding covariance matrix: , , , . In the first subsection, we study the effect of deleting each of the 20 observations on the two largest eigenvalues (which account for more than 99% of the total variation in principal component analysis) and on their corresponding eigenvectors. In the following subsection, we study the effects of deleting subsets of two observations on the largest eigenvalue. These subsets are those considered for the numerical study in Wang and Liski (1993). Finally, a short illustration of the potential effects of removing three observations is given in the last subsection.
Approximations When Deleting One Observation
Table 1 provides for each subset , the perturbed eigenvalue , its order one and two approximations and , the Rayleigh quotients ans and the differences between each approximation and the true perturbed eigenvalue . In the last two columns we find the bounds to and given by Inequality (12).
Table 1
Approximations and Error Analysis for the Largest Eigenvalue With r = 1
| I | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| 1 | 81.59976 | 81.8333 | 81.59769 | 81.58031 | 81.59969 | -0.23355 | 0.00207 | 0.01945 | 0.01988 | ||
| 2 | 77.18681 | 77.64932 | 77.18511 | 77.17611 | 77.18667 | -0.46251 | 0.00170 | 0.01070 | 0.01115 | ||
| 3 | 86.52309 | 86.52844 | 86.52306 | 86.52255 | 86.52309 | -0.00535 | |||||
| 4 | 77.34975 | 77.64477 | 77.32399 | 77.17133 | 77.34799 | -0.29503 | 0.02576 | 0.17842 | 0.18922 | ||
| 5 | 86.57878 | 86.58112 | 86.57875 | 86.57801 | 86.57878 | -0.00234 | |||||
| 6 | 76.85039 | 77.33877 | 76.8502 | 76.84922 | 76.85037 | -0.48838 | |||||
| 7 | 79.95562 | 80.27544 | 79.95368 | 79.94044 | 79.95552 | -0.31981 | 0.00194 | 0.01518 | 0.01544 | ||
| 8 | 74.71505 | 75.29919 | 74.71264 | 74.70229 | 74.71477 | -0.58414 | 0.00241 | 0.01276 | 0.01326 | ||
| 9 | 74.45165 | 75.05167 | 74.4498 | 74.44174 | 74.45144 | -0.60002 | 0.00185 | 0.01073 | |||
| 10 | 86.02449 | 86.05404 | 86.02442 | 86.02318 | 86.02449 | -0.02955 | |||||
| 11 | 85.32593 | 85.38314 | 85.3254 | 85.31697 | 85.32593 | -0.05721 | 0.001782881 | ||||
| 12 | 86.58382 | 86.58643 | 86.58381 | 86.58359 | 86.58382 | -0.00261 | |||||
| 13 | 84.13095 | 84.20755 | 84.12781 | 84.07951 | 84.13094 | -0.07660 | 0.00314 | 0.05143 | 0.05293 | ||
| 14 | 86.52446 | 86.52855 | 86.52439 | 86.52267 | 86.52446 | -0.00410 | |||||
| 15 | 82.72790 | 82.91768 | 82.72732 | 82.72175 | 82.72788 | -0.18978 | |||||
| 16 | 86.63917 | 86.63923 | 86.63917 | 86.63917 | 86.63917 | ||||||
| 17 | 83.27902 | 83.42004 | 83.27668 | 83.25055 | 83.27899 | -0.14101 | 0.00234 | 0.02847 | 0.02925 | ||
| 18 | 80.80479 | 81.08976 | 80.80396 | 80.79763 | 80.80476 | -0.28497 | |||||
| 19 | 78.33933 | 78.75217 | 78.33900 | 78.33701 | 78.33931 | -0.41283 | |||||
| 20 | 86.37150 | 86.38477 | 86.37149 | 86.37133 | 86.37150 | -0.01328 |
The first comment regarding the results in this table is that is by far the less accurate approximation in all cases. In contrast always provides extremely sharp approximations since it deviates from by in the worst case (when deleting Observation 4) and that the error is only when deleting Observation 16. It should be noted that also provides fairly satisfying approximations although clearly less accurate than . The Rayleigh quotient is outperformed by but remains significantly sharper than . Furthermore, it is worth pointing out that always overestimates the perturbed eigenvalue while the other three estimations slightly underestimate it. Note also that the results agree with inequalities and given in the Subsection Some Relations Between the Approximations.
Second, Inequality (12) provides bounds sufficiently close to and to evaluate correctly the accuracy of Rayleigh quotients as approximations to without having to recompute the perturbed analysis.
Third, it is easily seen from (10) that the maximum value for the perturbed eigenvalue is obtained when with . We have the highest perturbed eigenvalue when deleting Observations Number 3, 5, 10, 12, 14, 16, 20 which are fairly close to and in these cases we get the sharper approximations to .
Focusing now on the second largest eigenvalue, Table 2 provides results similar to those of Table 1. It turns out that is the less accurate approximation to . Except when deleting Observation 13, the Rayleigh quotient again provides the best approximation with a very good accuracy since in the worst case corresponding to this observation we have . For this observation, is slightly better but less accurate in all the other cases while performing fairly well in general. The Rayleigh quotient performs in a similar way as in Table 1.
Table 2
Approximations and Error Analysis for the Second Largest Eigenvalue with r = 1
| I | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| 1 | 6.80760 | 6.83824 | 6.80935 | 6.82480 | 6.80764 | -0.03065 | -0.00175 | -0.01721 | 0.22557 | ||
| 2 | 7.03262 | 7.04231 | 7.03372 | 7.03961 | 7.03271 | -0.00969 | -0.00110 | -0.00699 | 0.08423 | ||
| 3 | 6.73724 | 6.75509 | 6.73721 | 6.73727 | 6.73724 | -0.01785 | |||||
| 4 | 5.66955 | 5.87923 | 5.68474 | 5.81532 | 5.66891 | -0.20968 | -0.01519 | -0.14577 | 2.52388 | 0.02060 | |
| 5 | 6.10058 | 6.15035 | 6.10049 | 6.10071 | 6.10057 | -0.04977 | 0.01169 | ||||
| 6 | 7.09264 | 7.0928 | 7.09266 | 7.09276 | 7.09264 | ||||||
| 7 | 6.91733 | 6.94011 | 6.91920 | 6.93203 | 6.91743 | -0.02277 | -0.00187 | -0.01470 | 0.17207 | ||
| 8 | 7.03111 | 7.04295 | 7.03284 | 7.04029 | 7.03131 | -0.01184 | -0.00173 | -0.00917 | 0.09986 | ||
| 9 | 7.07835 | 7.08123 | 7.07877 | 7.080583 | 7.07840 | -0.00288 | 0.02615 | ||||
| 10 | 6.95776 | 6.96488 | 6.95776 | 6.95811 | 6.95776 | -0.00712 | 0.01375 | ||||
| 11 | 6.55595 | 6.59092 | 6.55642 | 6.56447 | 6.55595 | -0.03497 | -0.00851 | 0.11689 | |||
| 12 | 6.85444 | 6.86361 | 6.85422 | 6.85150 | 6.85444 | -0.00917 | 0.00293 | 0.00536 | |||
| 13 | 5.50884 | 5.62489 | 5.50826 | 5.54759 | 5.50807 | -0.11605 | -0.03875 | 0.84790 | |||
| 14 | 6.04911 | 6.07404 | 6.04375 | 6.02038 | 6.04857 | -0.02493 | 0.00536 | 0.02873 | 0.05689 | ||
| 15 | 6.98255 | 6.99306 | 6.98304 | 6.98777 | 6.98256 | -0.01051 | -0.00522 | 0.06610 | |||
| 16 | 7.08039 | 7.08097 | 7.08038 | 7.08031 | 7.08039 | ||||||
| 17 | 6.48050 | 6.53032 | 6.48147 | 6.50067 | 6.48043 | -0.04981 | -0.02017 | 0.3604578 | |||
| 18 | 7.04966 | 7.05459 | 7.04998 | 7.05253 | 7.04967 | -0.00492 | 0.03878 | ||||
| 19 | 7.09124 | 7.09159 | 7.09128 | 7.09148 | 7.09125 | ||||||
| 20 | 7.08664 | 7.08696 | 7.08664 | 7.08662 | 7.08664 |
It should be noted that again always overestimates the perturbed eigenvalue but in contrast to Table 1, the other three approximations can as well slightly underestimate or overestimate .
Another difference with Table 1 is that bounds provided to and by Inequality (12) are not so close to these quantities as they were previously. For a part, this can be explained by the fact that we have a smaller value for and than for and when considering the largest eigenvalue. Indeed, for the gap is smaller than the gap . However, if we except the case of Observation 4, we know from this bound that never exceeds and this is sufficient for practical interpretation.
Results for the eigenvectors corresponding to the two largest eigenvalues are given in Table 3 which provides the sines and for and their bounds provided by Inequality (13).
Table 3
Approximations and Error Analysis for the Eigenvectors Associated to the Two Largest Eigenvalues with r = 1
| I | ||||||||
|---|---|---|---|---|---|---|---|---|
| 1 | 0.01609 | 0.01644 | 0.01988 | 0.00153 | 0.19303 | 0.01056 | ||
| 2 | 0.01218 | 0.01267 | 0.01213 | 0.00142 | 0.11768 | 0.01250 | ||
| 3 | 0.00259 | 0.00261 | 0.00912 | 0.03408 | ||||
| 4 | 0.04974 | 0.05280 | 0.08114 | 0.01954 | 0.72248 | 0.06298 | ||
| 5 | 0.00309 | 0.00311 | 0.01097 | 0.00154 | 0.04563 | 0.00168 | ||
| 6 | 0.00393 | 0.00428 | 0.01491 | 0.00125 | 0.16533 | 0.01317 | ||
| 8 | 0.01356 | 0.01408 | 0.01310 | 0.00194 | 0.12507 | 0.01782 | ||
| 9 | 0.01169 | 0.01266 | 0.00725 | 0.00107 | 0.06553 | 0.00883 | ||
| 10 | 0.00403 | 0.00411 | 0.01092 | 0.04628 | ||||
| 11 | 0.01066 | 0.01078 | 0.01321 | 0.13947 | 0.00148 | |||
| 12 | 0.00167 | 0.00171 | 0.02195 | 0.02915 | ||||
| 13 | 0.02555 | 0.02630 | 0.05398 | 0.01230 | 0.42319 | 0.01285 | ||
| 14 | 0.00468 | 0.00477 | 0.07252 | 0.00970 | 0.10386 | 0.01011 | ||
| 15 | 0.00897 | 0.00913 | 0.01053 | 0.10247 | 0.00454 | |||
| 16 | 0.00335 | 0.00347 | ||||||
| 17 | 0.01918 | 0.01971 | 0.03624 | 0.00383 | 0.25215 | 0.00729 | ||
| 18 | 0.00962 | 0.01017 | 0.00999 | 0.07882 | 0.00486 | |||
| 19 | 0.00549 | 0.00597 | 0.00213 | 0.02021 | 0.00203 | |||
| 20 | 0.00142 | 0.00153 | 0.00274 | 0.00715 |
First, we consider the eigenvector associated to the largest eigenvalue. It could be noted that the maximum value of the sine bewteen the unperturbed and perturbed eigenvector is obtained when deleting Observation 4. This value corresponds to an angle of . In this case, as in all the other ones, the order one approximation performs fairly well since its sine with the perturbed eigenvector is equal to only 0.00494 which corresponds to an angle of . Furthermore bounds provided by Inequality (13) are always extremely close to the exact value of .
Focusing now on the eigenvector corresponding to the second largest eigenvalue, the maximum of is again obtained when deleting Observation 4 with the value of 0.08114. Even in this case, the order one approximation is fairly close to the perturbed eigenvector since . In contrast to the previous eigenvector, it is worth pointing out that bounds provided by Inequality (13) are not always sufficiently close to the true values of the sine to give an exact account of the accuracy for these approximations.
Finally, since the angles between the eigenvectors studied in the table are always very close to zero, we do not provide the tangent of these angles which only deviates from the sine by an extremely small amount.
Approximations When Deleting Subsets of Two Observations
Now, we study approximations to the perturbed largest eigenvalue ant its corresponding eigenvector when deleting the subsets of two observations considered for the numerical illustration in Wang and Liski (1993). Results similar to those of the previous subsection are provided in Tables 4 and 5.
Table 4
Approximations and Error Analysis for the Largest Eigenvalue with r = 2
| I | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| 64.59310 | 67.25005 | 64.58126 | 64.56084 | 64.59026 | -2.65695 | 0.01184 | 0.03226 | 0.00284 | 0.03616 | 0.00316 | |
| 67.51718 | 69.72595 | 67.4481 | 67.32138 | 67.50241 | -2.20877 | 0.06908 | 0.19580 | 0.01476 | 0.20842 | 0.01501 | |
| 67.98219 | 69.98722 | 67.85527 | 67.61002 | 67.95618 | -2.00502 | 0.12692 | 0.37218 | 0.02602 | 0.40664 | 0.02660 | |
| 69.40391 | 69.40294 | 69.40242 | 69.39558 | 69.40366 | 0.00149 | 0.00834 | 0.00892 | ||||
| 69.74526 | 69.73074 | 69.74115 | 69.72109 | 69.74455 | 0.01451 | 0.00410 | 0.02417 | 0.02512 | |||
| 69.98911 | 69.99202 | 69.98832 | 69.98402 | 69.98899 | -0.00291 | 0.00509 | 0.00565 | ||||
| 70.01929 | 72.14491 | 70.01294 | 69.99965 | 70.01808 | -2.12561 | 0.00636 | 0.01964 | 0.00121 | 0.02072 | 0.00126 | |
| 78.83587 | 79.36442 | 78.83242 | 78.82489 | 78.83542 | -0.52854 | 0.00345 | 0.01099 | 0.01144 | |||
| 72.60356 | 72.46791 | 72.58774 | 72.46791 | 72.60172 | 0.13565 | 0.01582 | 0.13565 | 0.00184 | 0.14414 | 0.00188 | |
| 75.53263 | 75.23992 | 75.50596 | 75.22661 | 75.53053 | 0.29271 | 0.02667 | 0.30602 | 0.00210 | 0.33028 | 0.00212 |
Table 5
Approximations and Error Analysis for the Eigenvector Associated to the Largest Eigenvalue with r = 2
| I | I | ||||||||
|---|---|---|---|---|---|---|---|---|---|
| 0.02254 | 0.00667 | 0.02525 | 0.00745 | 0.00856 | 0.00132 | 0.00951 | 0.00146 | ||
| 0.05637 | 0.01548 | 0.06010 | 0.01574 | 0.01740 | 0.00432 | 0.01833 | 0.00449 | ||
| 0.07724 | 0.02042 | 0.08464 | 0.02089 | 0.01220 | 0.00251 | 0.01269 | 0.00253 | ||
| 0.01128 | 0.00197 | 0.01205 | 0.00207 | 0.04477 | 0.00522 | 0.04759 | 0.00532 | ||
| 0.01950 | 0.00333 | 0.02026 | 0.00340 | 0.06609 | 0.00548 | 0.07148 | 0.00553 |
We note that deleting subsets of two observations can result in larger variations of the eigenvalue of interest than when deleting single observations. Indeed, the perturbed eigenvalue is lower than 70 in the first six lines of Table 4.
Furthermore for all the subsets I studied in this table we have a decrease of the eigenvalue, while we note that this eigenvalue is increased in several cases in Table 1. Despite these significant variations of the eigenvalue, we see that the Rayleigh quotient always provides a very accurate approximation to since the maximum gap observed for remains fairly moderate. It should also be noted that always performs better than . This point is fairly well illustrated considering again the case of for which we have . The Rayleigh quotient provides less accurate approximations than but should generally be preferred to if we except some cases. Finally, it is worth pointing out that bounds to and provided by Inequality (12) are always very close to the true values of these two differences thus avoiding to recompute the perturbed analysis.
Turning now to the sine values in Table 5 , we note the largest variations of the eigenvector when deleting the subsets , and . However, the order one approximation remains fairly satisfying in all the cases since the maximum value of obtained when deleting the subset does not exceed 0.02042 which corresponds to an angle of only .
Approximations When Deleting Subsets of Three Observations
While the accuracy of is fairly good for the removal of two observations despite the small sample size, one might think that results would probably be much worse when removing three observations and that their performance would deteriorate very quickly as increases. In order to investigate this point, we decided to perform a very brief study of the performance of the approximations for the largest eigenvalue when removing some sets of three observations. Note that in this case we have: instead of when deleting only two observations.
First, when removing the three Observations 3, 4 and 7, the largest eigenvalue initially equal to is moderately decreased to . The approximations to are then the following: , , and . We note the excellent accuracy of since the error is negligible.
Second, when removing the three Observations 2, 4 and 6, the largest eigenvalue is now significantly decreased to . The approximations to are then the following: , , and . We still have a satisfying accuracy of since the error is equal to 0.00295.
Finally, when removing the very influential set , the largest eigenvalue is drastically decreased to and we get the following approximations: , , and . Therefore while the largest eigenvalue is decreased by 26.939 from its initial value, the error remains fairly moderate while obviously larger than when considering removal of only two observations.
However, it would be necessary to perform additional studies with other data tables to get more reliable information on the relation between the value of ϵ and the accuracy of the approximations.
Concluding Remarks
The previous numerical study provides some indications on the sharpness of the various approximations in the paper when considering a particular data set. As a result, it may be useful to provide practitioners with some guidance on the choice of approximations for perturbed covariance matrices.
First, when focusing on eigenvalues, it should be noted that, in this example, approximations provided by should be avoided as well as the Rayleigh quotients which are not sufficiently accurate.
In contrast, Rayleigh quotients (based on the perturbed matrix and the approximations of order one ) seem to always provide reliable approximations to . They must be preferred to the approximation of order two in the study of this data table. Furthermore, their accuracy can be evaluated in a precise way by Inequality (12) if the eigenvalue of interest is not too close to the other eigenvalues, as emphasized by results in Table 1 when considering the largest eigenvalue. This is a definite advantage over other approximations.
It should also be noted that these Rayleigh quotients perform fairly well even for values of which are not necessarily very close to zero. This point is made clear in the first lines of Table 4 where and the eigenvalue is significantly decreased through the perturbation. The Subsection Approximations When Deleting Subsets of Three Observations, briefly dealing with the removal of three observations, tends to indicate that these Rayleigh quotients should still perform in a satisfying way even for higher values of ϵ. However, more in-depth studies would be necessary to get reliable conclusions on this point.
Second, when considering eigenvectors, we note a satisfying accuracy of approximations of order one . Again, when the eigenvalue corresponding to the eigenvector of interest is sufficiently distant from the other ones, we have a correct evaluation of this accuracy by Inequality (13).
This is an open access article distributed under the terms of the Creative Commons Attribution License (