scholarly journals Chain graph reduction into Power Chain Graphs

2021 ◽  
Author(s):  
Víthor Rosa Franco ◽  
Guilherme Wang Barros ◽  
Marie Wiberg ◽  
Jacob Arie Laros

Reduction of graphs is a class of procedures used to decrease the dimensionality of a given graph in which the properties of the reduced graph are to be induced from the properties of the larger original graph. This paper introduces both a new method for reducing chain graphs to simpler directed acyclic graphs (DAGs), that we call power chain graphs (PCG), as well as a procedure for structure learning of this new type of graph from correlational data of a Gaussian Graphical model (GGM). A definition for PCGs is given, directly followed by the reduction method. The structure learning procedure is a two-step approach: first, the correlation matrix is used to cluster the variables; and then, the averaged correlation matrix is used to discover the DAGs using the PC-stable algorithm. The results of simulations are provided to illustrate the theoretical proposal, which demonstrate initial evidence for the validity of our procedure to recover the structure of power chain graphs. The paper ends with a discussion regarding suggestions for future studies as well as some practical implications.

2018 ◽  
Vol 34 (2) ◽  
pp. 713-742
Author(s):  
Xiao Guo ◽  
Hai Zhang ◽  
Yao Wang ◽  
Yong Liang

Author(s):  
Mohammad Ali Javidian ◽  
Marco Valtorta ◽  
Pooyan Jamshidi

LWF chain graphs combine directed acyclic graphs and undirected graphs. We propose a PC-like algorithm, called PC4LWF, that finds the structure of chain graphs under the faithfulness assumption to resolve the problem of scalability of the proposed algorithm by Studeny (1997). We prove that PC4LWF is order dependent, in the sense that the output can depend on the order in which the variables are given. This order dependence can be very pronounced in high-dimensional settings. We propose two modifications of the PC4LWF algorithm that remove part or all of this order dependence. Simulation results with different sample sizes, network sizes, and p-values demonstrate the competitive performance of the PC4LWF algorithms in comparison with the LCD algorithm proposed by Ma et al. (2008) in low-dimensional settings and improved performance (with regard to error measures) in high-dimensional settings.


Entropy ◽  
2019 ◽  
Vol 21 (10) ◽  
pp. 975
Author(s):  
Aleksander Wieczorek ◽  
Volker Roth

Modelling causal relationships has become popular across various disciplines. Most common frameworks for causality are the Pearlian causal directed acyclic graphs (DAGs) and the Neyman-Rubin potential outcome framework. In this paper, we propose an information theoretic framework for causal effect quantification. To this end, we formulate a two step causal deduction procedure in the Pearl and Rubin frameworks and introduce its equivalent which uses information theoretic terms only. The first step of the procedure consists of ensuring no confounding or finding an adjustment set with directed information. In the second step, the causal effect is quantified. We subsequently unify previous definitions of directed information present in the literature and clarify the confusion surrounding them. We also motivate using chain graphs for directed information in time series and extend our approach to chain graphs. The proposed approach serves as a translation between causality modelling and information theory.


2013 ◽  
Vol 22 (02) ◽  
pp. 1350005 ◽  
Author(s):  
XIA LIU ◽  
YOULONG YANG ◽  
MINGMIN ZHU

Due to the infeasibility of randomized controlled experiments, the existence of unobserved variables and the fact that equivalent direct acyclic graphs obtained generally can not be distinguished, it is difficult to learn the true causal relations of original graph. This paper presents an algorithm called BSPC based on adjacent nodes to learn the structure of Causal Bayesian Networks with unobserved variables by using observational data. It does not have to adjust the structure as the existing algorithms FCI and MBCS*, while it can guarantee to obtain the true adjacent nodes. More important is that algorithm BSPC reduces computational complexity and improves reliability of conditional independence tests. Theoretical results show that the new algorithm is correct. In addition, the advantages of BSPC in terms of the number of conditional independence tests and the number of orientation errors are illustrated with simulation experiments from which we can see that it is more suitable in order to learn the structure of Causal Bayesian Networks with latent variables. Moreover a better latent structure representation is returned.


2019 ◽  
Vol 91 ◽  
pp. 78-87 ◽  
Author(s):  
Anna E. Austin ◽  
Tania A. Desrosiers ◽  
Meghan E. Shanahan

Sign in / Sign up

Export Citation Format

Share Document