Ouvi dizer que correlações parciais entre variáveis aleatórias podem ser encontradas invertendo a matriz de covariância e obtendo células apropriadas dessa matriz de precisão resultante (esse fato é mencionado em http://en.wikipedia.org/wiki/Partial_correlation , mas sem uma prova) .
Por que esse é o caso?
Respostas:
Quando uma variável aleatória multivariada(X1,X2,…,Xn) possui uma matriz de covariância não-regenerada C=(γij)=(Cov(Xi,Xj)) , o conjunto de todas as combinações lineares reais de Xi forma um n -dimensional espaço vectorial real com base E=(X1,X2,…,Xn) e um produto interno não degenerado, dado por
Sua base dupla em relação a esse produto interno , , é definida exclusivamente pelas relaçõesE∗=(X∗1,X∗2,…,X∗n)
the Kronecker delta (equal to1 when i=j and 0 otherwise).
The dual basis is of interest here because the partial correlation ofXi and Xj is obtained as the correlation between the part of Xi that is left after projecting it into the space spanned by all the other vectors (let's simply call it its "residual", Xi∘ ) and the comparable part of Xj , its residual Xj∘ . Yet X∗i is a vector that is orthogonal to all vectors besides Xi and has positive inner product with Xi whence Xi∘ must be some non-negative multiple of X∗i , and likewise for Xj . Let us therefore write
for positive real numbersλi and λj .
The partial correlation is the normalized dot product of the residuals, which is unchanged by rescaling:
(In either case the partial correlation will be zero whenever the residuals are orthogonal, whether or not they are nonzero.)
We need to find the inner products of dual basis elements. To this end, expand the dual basis elements in terms of the original basisE :
Then by definition
In matrix notation withI=(δij) the identity matrix and B=(βij) the change-of-basis matrix, this states
That is,B=C−1 , which is exactly what the Wikipedia article is asserting. The previous formula for the partial correlation gives
fonte
Here is a proof with just matrix calculations.
I appreciate the answer by whuber. It is very insightful on the math behind the scene. However, it is still not so trivial how to use his answer to obtain the minus sign in the formula stated in the wikipediaPartial_correlation#Using_matrix_inversion.
To get this minus sign, here is a different proof I found in "Graphical Models Lauriten 1995 Page 130". It is simply done by some matrix calculations.
The key is the following matrix identity:
Write down the covariance matrix as
LetP=Ω−1 . Similarly, write down P as
By the key matrix identity,
We also know thatΩ11−Ω12Ω−122Ω21 is the covariance matrix of (Xi,Xj)|V∖{Xi,Xj} (from Multivariate_normal_distribution#Conditional_distributions). The partial correlation is therefore
Just simple inversion formula of 2-by-2 matrix,
Therefore,
fonte
i=j
, thenrho_ii V\{X_i, X_i} = -1
, How do we interpret those diagonal elements in the precision matrix?Note that the sign of the answer actually depends on how you define partial correlation. There is a difference between regressingXi and Xj on the other n−1 variables separately vs. regressing Xi and Xj on the other n−2 variables together. Under the second definition, let the correlation between residuals ϵi and ϵj be ρ . Then the partial correlation of the two (regressing ϵi on ϵj and vice versa) is −ρ .
This explains the confusion in the comments above, as well as on Wikipedia. The second definition is used universally from what I can tell, so there should be a negative sign.
I originally posted an edit to the other answer, but made a mistake - sorry about that!
fonte