scholarly journals Parameter Learning of Logic Programs for Symbolic-Statistical Modeling

2001 ◽  
Vol 15 ◽  
pp. 391-454 ◽  
Author(s):  
T. Sato ◽  
Y. Kameya

We propose a logical/mathematical framework for statistical parameter learning of parameterized logic programs, i.e. definite clause programs containing probabilistic facts with a parameterized distribution. It extends the traditional least Herbrand model semantics in logic programming to distribution semantics, possible world semantics with a probability distribution which is unconditionally applicable to arbitrary logic programs including ones for HMMs, PCFGs and Bayesian networks. We also propose a new EM algorithm, the graphical EM algorithm, that runs for a class of parameterized logic programs representing sequential decision processes where each decision is exclusive and independent. It runs on a new data structure called support graphs describing the logical relationship between observations and their explanations, and learns parameters by computing inside and outside probability generalized for logic programs. The complexity analysis shows that when combined with OLDT search for all explanations for observations, the graphical EM algorithm, despite its generality, has the same time complexity as existing EM algorithms, i.e. the Baum-Welch algorithm for HMMs, the Inside-Outside algorithm for PCFGs, and the one for singly connected Bayesian networks that have been developed independently in each research field. Learning experiments with PCFGs using two corpora of moderate size indicate that the graphical EM algorithm can significantly outperform the Inside-Outside algorithm.

Author(s):  
Ming-Sheng Ying ◽  
Yuan Feng ◽  
Sheng-Gang Ying

AbstractMarkov decision process (MDP) offers a general framework for modelling sequential decision making where outcomes are random. In particular, it serves as a mathematical framework for reinforcement learning. This paper introduces an extension of MDP, namely quantum MDP (qMDP), that can serve as a mathematical model of decision making about quantum systems. We develop dynamic programming algorithms for policy evaluation and finding optimal policies for qMDPs in the case of finite-horizon. The results obtained in this paper provide some useful mathematical tools for reinforcement learning techniques applied to the quantum world.


2012 ◽  
Vol 532-533 ◽  
pp. 1445-1449
Author(s):  
Ting Ting Tong ◽  
Zhen Hua Wu

EM algorithm is a common method to solve mixed model parameters in statistical classification of remote sensing image. The EM algorithm based on fuzzification is presented in this paper to use a fuzzy set to represent each training sample. Via the weighted degree of membership, different samples will be of different effect during iteration to decrease the impact of noise on parameter learning and to increase the convergence rate of algorithm. The function and accuracy of classification of image data can be completed preferably.


2017 ◽  
Vol 4 (11) ◽  
pp. 171377 ◽  
Author(s):  
Xiaoguang Huo ◽  
Feng Fu

Sequential portfolio selection has attracted increasing interest in the machine learning and quantitative finance communities in recent years. As a mathematical framework for reinforcement learning policies, the stochastic multi-armed bandit problem addresses the primary difficulty in sequential decision-making under uncertainty, namely the exploration versus exploitation dilemma, and therefore provides a natural connection to portfolio selection. In this paper, we incorporate risk awareness into the classic multi-armed bandit setting and introduce an algorithm to construct portfolio. Through filtering assets based on the topological structure of the financial market and combining the optimal multi-armed bandit policy with the minimization of a coherent risk measure, we achieve a balance between risk and return.


Author(s):  
R. Tse ◽  
G. Seet ◽  
S. K. Sim

Controlling a robot to perform a task is more difficult than commanding a human. A robot needs to be preprogrammed to perform a task. This is achieved by providing the robot with a complete set of step-by-step commands from the beginning till the end. In contrast, to a human, recalling an experience when he was instructed with the same command in a similar situation, a human would be able to guess what intention behind such a command is and could then behave cooperatively. Our objective is to equip the robot with such a capability of recognizing some simple human intentions required of a robot, such as: moving around a corner, moving parallel to the wall, or moving towards an object. The cues used by the robot to make an inference were: the odometer and laser sensor readings, and the human operator’s commands given. Using the Maximum-Likelihood (ML) parameter learning on Dynamic Bayesian Networks, the correlations between these cues and the intentions were modeled and used to infer the human intentions in controlling the robot. From the experiments, the robot was able to learn and infer the above mentioned intentions of the human user with a satisfying success rate.


Sign in / Sign up

Export Citation Format

Share Document