Theory.Rmd
Let \(X = (X_1, ..., X_p)\) be a random vector and suppose the distribution of \(X\) admits to a linear structual equation model (SEM), that is \[X_j = \sum_{k\neq j}\beta_{jk}X_k + \varepsilon_j, \quad \text{for } j = 1, ..., p,\] where \(\varepsilon_1, ...,\varepsilon_p\) are independent, zero mean random variables, and \((\beta_{jk})_{j,k\in \{1,...,p\}}\) are unknown regression parameters. Assume furthermore that that all \(\varepsilon_j\) have a common and unknown finite variance \(\sigma^2>0\). The covariance matrix of \(X\) may be calculated by \[\Sigma = \mathbb{E}(XX^T)=\sigma^2(I-B)^{-1}(I-B)^{-T},\] where \(B\) is a \(p\times p\) matrix with \(jk\)th entry equal to \(\beta_{jk}\) and zeros in the diagonal
This SEM induces a directed graph with verticies \(V=\{1, ..., p\}\) and edges \(E=\{(j,k): \beta_{jk} \neq 0\}\). We will assume that this graph is acyclic, which ensures that the graph has at least one sink and one source node.
In Chen, Drton and Wang (2018) it is shown that
The first two points allows us to identify source nodes in the graph by finding the variables with minimal variance in \(\Sigma\), and the last two points allows us to identify sink nodes in the graph by finding the variables with minimal precision in \(\Sigma^{-1}\).
For an index \(j\in \{1,...,p\}\) and set of indicies \(C\subseteq \{1, ..., p\}\) define \[X_{j.C} = X_j - \mathbb{E}[X_j\mid (X_k: k\in C)].\] Suppose now that \(C\) is an ansestral set, that is, for all \(j\in C\) it holds that \(an(j)\subseteq C\). Then for all \(j\in C\) it holds that \(X_{j.C}=0\) and \[X_j = \sum_{k\in C}\beta_{jk}X_k + \varepsilon_j,\] furthermore, for all \(j\not\in C\) it holds that \[X_{j.C} = \sum_{k\in pa(j)\setminus C}\beta_{jk}X_{k.C} + \varepsilon_j\]
Combining these results with the points above we are able to identify the topological ordering of the graph \(\mathcal{G}\) by one of the the two following methods