Introduction and Historical Background

This chapter describes the evolution of different multi-objective decision-making (MODM) models with their historical backgrounds. Starting from MODM models in deterministic environments along with various solution techniques, the chapter presents how different kinds of uncertainties may be associated with such decision-making models. Among several types of uncertainties, it has been found that probabilistic and possibilistic uncertainties are of special interests. A brief literature survey on different existing methods to solve those types of uncertainties, independently, is discussed and focuses on the need of considering simultaneous occurrence of those types of uncertainties in MODM contexts. Finally, a bibliographic survey on several approaches for MODM under hybrid fuzzy environments has been presented. Through this chapter the readers can be able to get some concepts about the historical development of MODM models in hybrid fuzzy environments and their importance in solving various real-life problems in the current complex decision-making arena.

2021 ◽  
Author(s):  
Rujing Zha ◽  
Peng Li ◽  
Ying Li ◽  
Nan Li ◽  
Meijun Gao ◽  
...  

Abstract A good-based model proposes that the orbitofrontal cortex (OFC) represents binary choice outcome, i.e., the chosen good. Previous studies have found that the OFC represents the binary choice outcome in decision-making tasks involving commodity type, cost, risk, and delay. Real-life decisions are often complex and involve uncertainty, rewards, and penalties; however, whether the OFC represents binary choice outcomes in a such decision-making situation, e.g., Iowa gambling task (IGT), remains unclear. Here, we propose that the OFC represents binary choice outcome, i.e., advantageous choice versus disadvantageous choice, in the IGT. We propose two hypotheses: first, the activity pattern in the human OFC represents an advantageous choice; and second, choice induces an OFC-related functional network. Using functional magnetic resonance imaging and advanced machine learning tools, we found that the OFC represented an advantageous choice in the IGT. The OFC representation of advantageous choice was related to decision-making performance. Choice modulated the functional connectivity between the OFC and the superior medial gyrus. In conclusion, the OFC represents an advantageous choice during the IGT. In the framework of a good-based model, the results extend the role of the OFC to complex decision-making when making a binary choice.


Mathematics ◽  
2021 ◽  
Vol 9 (13) ◽  
pp. 1456
Author(s):  
Stefka Fidanova ◽  
Krassimir Todorov Atanassov

Some of industrial and real life problems are difficult to be solved by traditional methods, because they need exponential number of calculations. As an example, we can mention decision-making problems. They can be defined as optimization problems. Ant Colony Optimization (ACO) is between the best methods, that solves combinatorial optimization problems. The method mimics behavior of the ants in the nature, when they look for a food. One of the algorithm parameters is called pheromone, and it is updated every iteration according quality of the achieved solutions. The intuitionistic fuzzy (propositional) logic was introduced as an extension of Zadeh’s fuzzy logic. In it, each proposition is estimated by two values: degree of validity and degree of non-validity. In this paper, we propose two variants of intuitionistic fuzzy pheromone updating. We apply our ideas on Multiple-Constraint Knapsack Problem (MKP) and compare achieved results with traditional ACO.


2021 ◽  
Vol 35 (2) ◽  
Author(s):  
Nicolas Bougie ◽  
Ryutaro Ichise

AbstractDeep reinforcement learning methods have achieved significant successes in complex decision-making problems. In fact, they traditionally rely on well-designed extrinsic rewards, which limits their applicability to many real-world tasks where rewards are naturally sparse. While cloning behaviors provided by an expert is a promising approach to the exploration problem, learning from a fixed set of demonstrations may be impracticable due to lack of state coverage or distribution mismatch—when the learner’s goal deviates from the demonstrated behaviors. Besides, we are interested in learning how to reach a wide range of goals from the same set of demonstrations. In this work we propose a novel goal-conditioned method that leverages very small sets of goal-driven demonstrations to massively accelerate the learning process. Crucially, we introduce the concept of active goal-driven demonstrations to query the demonstrator only in hard-to-learn and uncertain regions of the state space. We further present a strategy for prioritizing sampling of goals where the disagreement between the expert and the policy is maximized. We evaluate our method on a variety of benchmark environments from the Mujoco domain. Experimental results show that our method outperforms prior imitation learning approaches in most of the tasks in terms of exploration efficiency and average scores.


Sign in / Sign up

Export Citation Format

Share Document