ランダム変数間の偏相関は、共分散行列を反転し、そのような結果の精度行列から適切なセルを取得することで見つけることができると聞きました(この事実は http://en.wikipedia.org/wiki/Partial_correlationにいますが、証拠はありません) 。
これはなぜですか?
ランダム変数間の偏相関は、共分散行列を反転し、そのような結果の精度行列から適切なセルを取得することで見つけることができると聞きました(この事実は http://en.wikipedia.org/wiki/Partial_correlationにいますが、証拠はありません) 。
これはなぜですか?
回答:
場合、多変量確率変数持つ非縮退共分散行列C = (γ I 、J)= (Covを(X I、XのJ))、すべての実際の線形結合の組X iは、基底E = (X 1、X 2、… でn次元の実ベクトル空間を形成します。および非縮退内積
その双対基底この内積に対して、、一意の関係によって定義されます
the Kronecker delta (equal to when and otherwise).
The dual basis is of interest here because the partial correlation of and is obtained as the correlation between the part of that is left after projecting it into the space spanned by all the other vectors (let's simply call it its "residual", ) and the comparable part of , its residual . Yet is a vector that is orthogonal to all vectors besides and has positive inner product with whence must be some non-negative multiple of , and likewise for . Let us therefore write
for positive real numbers and .
The partial correlation is the normalized dot product of the residuals, which is unchanged by rescaling:
(In either case the partial correlation will be zero whenever the residuals are orthogonal, whether or not they are nonzero.)
We need to find the inner products of dual basis elements. To this end, expand the dual basis elements in terms of the original basis :
Then by definition
In matrix notation with the identity matrix and the change-of-basis matrix, this states
That is, , which is exactly what the Wikipedia article is asserting. The previous formula for the partial correlation gives
Here is a proof with just matrix calculations.
I appreciate the answer by whuber. It is very insightful on the math behind the scene. However, it is still not so trivial how to use his answer to obtain the minus sign in the formula stated in the wikipediaPartial_correlation#Using_matrix_inversion.
To get this minus sign, here is a different proof I found in "Graphical Models Lauriten 1995 Page 130". It is simply done by some matrix calculations.
The key is the following matrix identity:
Write down the covariance matrix as
Let . Similarly, write down as
By the key matrix identity,
We also know that is the covariance matrix of (from Multivariate_normal_distribution#Conditional_distributions). The partial correlation is therefore
Just simple inversion formula of 2-by-2 matrix,
Therefore,
i=j
, then rho_ii V\{X_i, X_i} = -1
, How do we interpret those diagonal elements in the precision matrix?
Note that the sign of the answer actually depends on how you define partial correlation. There is a difference between regressing and on the other variables separately vs. regressing and on the other variables together. Under the second definition, let the correlation between residuals and be . Then the partial correlation of the two (regressing on and vice versa) is .
This explains the confusion in the comments above, as well as on Wikipedia. The second definition is used universally from what I can tell, so there should be a negative sign.
I originally posted an edit to the other answer, but made a mistake - sorry about that!