scholarly journals AND/OR Multi-Valued Decision Diagrams (AOMDDs) for Graphical Models

2008 ◽  
Vol 33 ◽  
pp. 465-519 ◽  
Author(s):  
R. Mateescu ◽  
R. Dechter ◽  
R. Marinescu

Inspired by the recently introduced framework of AND/OR search spaces for graphical models, we propose to augment Multi-Valued Decision Diagrams (MDD) with AND nodes, in order to capture function decomposition structure and to extend these compiled data structures to general weighted graphical models (e.g., probabilistic models). We present the AND/OR Multi-Valued Decision Diagram (AOMDD) which compiles a graphical model into a canonical form that supports polynomial (e.g., solution counting, belief updating) or constant time (e.g. equivalence of graphical models) queries. We provide two algorithms for compiling the AOMDD of a graphical model. The first is search-based, and works by applying reduction rules to the trace of the memory intensive AND/OR search algorithm. The second is inference-based and uses a Bucket Elimination schedule to combine the AOMDDs of the input functions via the the APPLY operator. For both algorithms, the compilation time and the size of the AOMDD are, in the worst case, exponential in the treewidth of the graphical model, rather than pathwidth as is known for ordered binary decision diagrams (OBDDs). We introduce the concept of semantic treewidth, which helps explain why the size of a decision diagram is often much smaller than the worst case bound. We provide an experimental evaluation that demonstrates the potential of AOMDDs.

2017 ◽  
Vol 42 (2) ◽  
Author(s):  
Vilda Purutçuoğlu ◽  
Ezgi Ayyıldız ◽  
Ernst Wit

AbstractIntroduction:The Gaussian Graphical Model (GGM) is one of the well-known probabilistic models which is based on the conditional independency of nodes in the biological system. Here, we compare the estimates of the GGM parameters by the graphical lasso (glasso) method and the threshold gradient descent (TGD) algorithm.Methods:We evaluate the performance of both techniques via certain measures such as specificity, F-measure and AUC (area under the curve). The analyses are conducted by Monte Carlo runs under different dimensional systems.Results:The results indicate that the TGD algorithm is more accurate than the glasso method in all selected criteria, whereas, it is more computationally demanding than this method too.Discussion and conclusion:Therefore, in high dimensional systems, we recommend glasso for its computational efficiency in spite of its loss in accuracy and we believe than the computational cost of the TGD algorithm can be improved by suggesting alternative steps in inference of the network.


Biometrika ◽  
2020 ◽  
Author(s):  
S Na ◽  
M Kolar ◽  
O Koyejo

Abstract Differential graphical models are designed to represent the difference between the conditional dependence structures of two groups, thus are of particular interest for scientific investigation. Motivated by modern applications, this manuscript considers an extended setting where each group is generated by a latent variable Gaussian graphical model. Due to the existence of latent factors, the differential network is decomposed into sparse and low-rank components, both of which are symmetric indefinite matrices. We estimate these two components simultaneously using a two-stage procedure: (i) an initialization stage, which computes a simple, consistent estimator, and (ii) a convergence stage, implemented using a projected alternating gradient descent algorithm applied to a nonconvex objective, initialized using the output of the first stage. We prove that given the initialization, the estimator converges linearly with a nontrivial, minimax optimal statistical error. Experiments on synthetic and real data illustrate that the proposed nonconvex procedure outperforms existing methods.


1993 ◽  
Vol 03 (01) ◽  
pp. 3-12 ◽  
Author(s):  
DETLEF SIELING ◽  
INGO WEGENER

(Ordered) binary decision diagrams are a powerful representation for Boolean functions and are widely used in logical synthesis, verification, test pattern generation or as part of CAD tools. NC-algorithms are presented for the most important operations on this representation, e.g. evaluation for a given input, minimization, satisfiability, redundancy test, replacement of variables by constants or functions, equivalence test and synthesis. The algorithms have logarithmic run time on CRCW COMMON PRAMs with a polynomial number of processors.


2000 ◽  
Vol 103 (1-3) ◽  
pp. 237-258 ◽  
Author(s):  
Martin Sauerhoff ◽  
Ingo Wegener ◽  
Ralph Werchner

2010 ◽  
Vol 5 (1) ◽  
pp. 78-88 ◽  
Author(s):  
Marcelo Porto ◽  
André Silva ◽  
Sergo Almeida ◽  
Eduardo Da Costa ◽  
Sergio Bampi

This paper presents real time HDTV (High Definition Television) architecture for Motion Estimation (ME) using efficient adder compressors. The architecture is based on the Quarter Sub-sampled Diamond Search algorithm (QSDS) with Dynamic Iteration Control (DIC) algorithm. The main characteristic of the proposed architecture is the large amount of Processing Units (PUs) that are used to calculate the SAD (Sum of Absolute Difference) metric. The internal structures of the PUs are composed by a large number of addition operations to calculate the SADs. In this paper, efficient 4-2 and 8-2 adder compressors are used in the PUs architecture to achieve the performance to work with HDTV (High Definition Television) videos in real time at 30 frames per second. These adder compressors enable the simultaneous addition of 4 and 8 operands respectively. The PUs, using adder compressors, were applied to the ME architecture. The implemented architecture was described in VHDL and synthesized to FPGA and, with Leonardo Spectrum tool, to the TSMC 0.18μm CMOS standard cell technology. Synthesis results indicate that the new QSDS-DIC architecture reach the best performance result and enable gains of 12% in terms of processing rate. The architecture can reach real time for full HDTV (1920x1080 pixels) in the worst case processing 65 frames per second, and it can process 269 HDTV frames per second in the average case.


2006 ◽  
Vol 6 (6) ◽  
pp. 483-494
Author(s):  
T. Tulsi ◽  
L.K. Grover ◽  
A. Patel

The standard quantum search lacks a feature, enjoyed by many classical algorithms, of having a fixed point, i.e. monotonic convergence towards the solution. Recently a fixed point quantum search algorithm has been discovered, referred to as the Phase-\pi/3 search algorithm, which gets around this limitation. While searching a database for a target state, this algorithm reduces the error probability from \epsilon to \epsilon^{2q+1} using q oracle queries, which has since been proved to be asymptotically optimal. A different algorithm is presented here, which has the same worst-case behavior as the Phase-\pi/3 search algorithm but much better average-case behavior. Furthermore the new algorithm gives \epsilon^{2q+1} convergence for all integral q, whereas the Phase-\pi/3 search algorithm requires q to be (3^{n}-1)/2 with n a positive integer. In the new algorithm, the operations are controlled by two ancilla qubits, and fixed point behavior is achieved by irreversible measurement operations applied to these ancillas. It is an example of how measurement can allow us to bypass some restrictions imposed by unitarity on quantum computing.


2021 ◽  
Vol 3 (1) ◽  
pp. 40-48
Author(s):  
Sivaganesan D

A network of tiny sensors located at various regions for sensing and transmitting information is termed as wireless sensor networks. The information from multiple network nodes reach the destination node or the base station where data processing is performed. In larger search spaces, the clustering mechanisms and routing solutions provided by the existing heuristic algorithms are often inefficient. The sensor node resources are depleted by un-optimized processes created by reduced routing and clustering optimization levels in large search spaces. Chaotic Gravitational Search Algorithm and Fuzzy based clustering schemes are used to overcome the limitations and challenges of the conventional routing systems. This enables effective routing and efficient clustering in large search spaces. In each cluster, among the available nodes, appropriate node is selected as the cluster head. Reduction in delay, increase in energy consumption, increase in network lifetime and improvement of the network clustering accuracy are evident from the simulation results.


Sign in / Sign up

Export Citation Format

Share Document