Reasoning Without the Conjunction Closure

Episteme ◽  
2021 ◽  
pp. 1-14
Author(s):  
Alicja Kowalewska

Abstract Some theories of rational belief assume that beliefs should be closed under conjunction. I motivate the rejection of the conjunction closure, and point out that the consequences of this rejection are not as severe as it is usually thought. An often raised objection is that without the conjunction closure people are unable to reason. I outline an approach in which we can – in usual cases – reason using conjunctions without accepting the closure in its whole generality. This solution is based on the notion of confidence levels, which can be defined using probabilities. Moreover, on this approach, reasoning has a scalable computational complexity adaptable to cognitive abilities of both rationally bounded and perfectly rational agents. I perform a simulation to assess its error rate, and compare it to reasoning with conjunction closure.

Episteme ◽  
2019 ◽  
pp. 1-20
Author(s):  
Eyal Tal

ABSTRACTShould conciliating with disagreeing peers be considered sufficient for reaching rational beliefs? Thomas Kelly argues that when taken this way, Conciliationism lets those who enter into a disagreement with an irrational belief reach a rational belief all too easily. Three kinds of responses defending Conciliationism are found in the literature. One response has it that conciliation is required only of agents who have a rational belief as they enter into a disagreement. This response yields a requirement that no one should follow. If the need to conciliate applies only to already rational agents, then an agent must conciliate only when her peer is the one irrational. A second response views conciliation as merely necessary for having a rational belief. This alone does little to address the central question of what is rational to believe when facing a disagreeing peer. Attempts to develop the response either reduce to the first response, or deem necessary an unnecessary doxastic revision, or imply that rational dilemmas obtain in cases where intuitively there are none. A third response tells us to weigh what our pre-disagreement evidence supports against the evidence from the disagreement itself. This invites epistemic akrasia.


Episteme ◽  
2016 ◽  
Vol 14 (4) ◽  
pp. 519-538 ◽  
Author(s):  
Robert Mark simpson

ABSTRACTPermissivism says that for some propositions and bodies of evidence, there is more than one rationally permissible doxastic attitude that can be taken towards that proposition given the evidence. Some critics of this view argue that it condones, as rationally acceptable, sets of attitudes that manifest an untenable kind of arbitrariness. I begin by providing a new and more detailed explication of what this alleged arbitrariness consists in. I then explain why Miriam Schoenfield's prima facie promising attempt to answer the Arbitrariness Objection, by appealing to the role of epistemic standards in rational belief formation, fails to resolve the problem. Schoenfield's strategy is, however, a useful one, and I go on to explain how an alternative form of the standards-based approach to Permissivism – one that emphasizes the significance of the relationship between people's cognitive abilities and the epistemic standards that they employ – can respond to the arbitrariness objection.


2013 ◽  
Vol 760-762 ◽  
pp. 1684-1689
Author(s):  
Hong Hao Zhao ◽  
Fan Bo Meng ◽  
Qing Qi Zhao ◽  
Jie Wang

Recently, structured LDPC codes have been focused on due to excellent performance and lower complexity. An improved construction of QC-LDPC based on a modified PEG algorithm is proposed in this paper. The modified Progressive Edge-Growth algorithm is a PEG algorithm with Approximated Cycle Extrinsic Message Degree (ACE) metric, which is used to describe the connectivity of cycles. The approach can maximize the girth of the cycles, improve the connectivity of cycles and have the advantages of QC algorithms. The simulation results demonstrate that the PEG algorithm based on ACE has lower Bit Error Rate (BER) and Frame Error Rate (FER) than the original PEG algorithm at the low Signal Noise Ratio (SNR) values, and has a relatively lower computational complexity.


2021 ◽  
Vol 22 (1) ◽  
pp. 201-262
Author(s):  
Mark Spottswood

Abstract This Article explores settlement incentives under three different burden of proof rules. The conventional burden of proof is a discontinuous step-function, jumping from no damages to full damages at the 0.5 jury confidence level. Continuous burdens of proof, by contrast, would permit sanctions to steadily increase as juror confidence rises from 0 to 1, with no discontinuity. Linear burdens, which have received extensive attention in prior literature, escalate sanctions steadily across the whole range of confidence levels, while the logistic burden takes a nonlinear form. Using a data simulation approach guided by the empirical realities of American civil litigation, I consider the incentives that each of these rules creates for parties contemplating settlement, using a model in which parties make divergent forecasts of their expected outcomes at trial due to optimism bias. Based on this analysis, I conclude that a linear burden would likely raise our settlement rate by a modest amount, except in very large cases and in “easy” cases, in which an unbiased person would predict that a trial factfinder would have a level of confidence in liability quite close to either zero or one. I also compare the expected error rate of the settlements that each rule produces, and find that the linear rule modestly lowers the expected error rate of settlement overall, although this benefit does not hold for easy cases or those with very high damages. Lastly, I conduct a similar analysis for the logistic burden, finding that it induces a similar quality and quantity of settlements as we currently achieve using conventional burdens.


2020 ◽  
Vol 17 (1) ◽  
pp. 254-259
Author(s):  
Harikrishna Ponnam ◽  
Jakeer Hussain Shaik

In the application of remote cardiovascular monitoring, the computational complexity and power consumption need to be maintained in a considerable level in order to prevent the limitations introduced by the computationally constrained equipment’s that perform the process of continuous monitoring and analysis. In this paper, a Circulant Matrix-based Continuous Wavelet Transform (CM-CWT)-based feature extraction mechanism is contributed to minimizing the computational complexity incurred during the process of feature extraction from the input ECG signals. This proposed CM-CWT mechanism derives the advantages of the Circulant Matrix-based Continuous Wavelet Transform and Gradient-based filtering design for achieving excellent feature extraction from ECG signals with low computational complexity. The experimental investigation of the proposed CM-CWT mechanism is conducted using the factors of computational complexity, sensitivity, prediction accuracy and error rate for estimating its predominance over the compared DWT-HAAR and HIFEA approaches used for ECG feature extraction. The experiments of the proposed CM-CWT mechanism on an average is estimated to reduce the error rate to the maximum of 21% compared to the existing DWT-HAAR and HIFEA approaches used for ECG feature extraction.


1986 ◽  
Vol 80 (4) ◽  
pp. 1187-1207 ◽  
Author(s):  
Jonathan Bendor ◽  
Terry M. Moe

Most models of agenda control examine dyadic relations—for example, those between a committee and the floor of a legislature. Such relations, however, are always embedded in a larger context, namely, a political environment composed of voters and interest groups. In this paper we model agenda setters (a legislative committee) as decision makers with limited cognitive abilities who adjust over time to their larger political environment. The legislators' policy positions are endogenous, reflecting the relative strengths of voters wielding the district-specific resource of votes and of interest groups wielding the transferable resource of money. The resulting outcomes indicate that neoclassical models of voting and pluralist models of group influence have each told part of the story. When only votes matter, our boundedly rational agents grope toward equilibria close to those of neoclassical models; however, when mobile resources matter as well, the outcomes depart systematically from those of previous models. In particular, interest groups can make themselves worse off by capturing the committee. The results suggest that agenda control is less powerful than conventionally believed and point toward conditions shaping its effectiveness—conditions highlighting the distinctive contributions of pluralist and neoclassical thinking to a broader theory of political institutions.


2011 ◽  
Vol 271-273 ◽  
pp. 1059-1062
Author(s):  
Lei Wang ◽  
Bao Yu Zheng ◽  
Jing Wu Cui

Based on the synthesis and analysis of recursive receivers, new algorithms are proposed to achieve satisfactory performance with moderate computational complexity. During the analysis, some interesting properties shared by the proposed procedures are described. Finally, the performance assessment shows that new schemes are superior to the linear detector and ordinary grouping algorithm, and achieve a bit-error rate close to that of the optimum receiver.


Author(s):  
Jody Azzouni

The word “know” is revealed as vague, applicable to fallible agents, factive, and criterion-transcendent. It is invariant in its meaning across contexts and invariant relative to different agents. Only purely epistemic properties affect its correct application—not the interests of agents or those who attribute the word to agents. These properties enable “know” to be applied correctly—as it routinely is—to cognitive agents ranging from sophisticated human knowers, who engage in substantial metacognition, to various animals, who know much less and do much less, if any, metacognition, to nonconscious mechanical devices such as drones, robots, and the like. These properties of the word “know” suffice to explain the usage phenomena that contextualists and subject-sensitive invariantists invoke to place pressure on an understanding of the word that treats its application as involving no interests of agents, or others. It is also shown that the factivity and the fallibilist-compatibility of the word “know” explain Moorean paradoxes, the preface paradox, and the lottery paradox. A fallibility-sensitive failure of knowledge closure is given along with a similar failure of rational-belief closure. The latter explains why rational agents can nevertheless believe A and B, where A and B contradict each other. A substantial discussion of various kinds of metacognition is given—as well as a discussion of the metacognition literature in cognitive ethology. An appendix offers a new resolution of the hangman paradox, one that turns neither on a failure of knowledge closure nor on a failure of KK.


2021 ◽  
Vol 11 (2) ◽  
pp. 24
Author(s):  
Guobin Xia ◽  
Muzi Li ◽  
Philip Henry ◽  
Stephen Westland ◽  
Francisco Queiroz ◽  
...  

The purpose of this study was to explore the influence of environmental colour on people’s lateral and logical abilities. This was done by evaluating study participants’ response time and error rate when completing six types of psychometric tests that were performed in various hue backgrounds on a computer. To maximise the colour stimulation provided by the monitor, the experiment was carried out in a dark laboratory. Analysis of participants’ response time and error rate showed that different colours could significantly influence arousal and impulsiveness, which suggests that colour has indirect impacts on cognitive abilities. Further analysis revealed that different colours had various effects depending on the type of psychometric test given. These findings suggest that future research on environmental design should consider how to effectively use colour to impact people’s performance and behaviour.


2021 ◽  
Author(s):  
Fabio Aurelio D'Asaro ◽  
Paolo Baldi ◽  
Giuseppe Primiero

Depth-Bounded Boolean Logics (DBBL for short) are well-understood frameworks to model rational agents equipped with limited deductive capabilities. These Logics use a parameter k>=0 to limit the amount of virtual information, i.e., the information that the agent may temporarily assume throughout the deductive process. This restriction brings several advantageous properties over classical Propositional Logic, including polynomial decision procedures for deducibility and refutability. Inspired by DBBL, we propose a limited-depth version of the popular ASP system \clingo, tentatively dubbed k-lingo after the bound k on virtual information. We illustrate the connection between DBBL and ASP through examples involving both proof-theoretical and implementative aspects. The paper concludes with some comments on future work, which include a computational complexity characterization of the system, applications to multi-agent systems and feasible approximations of probability functions.


Sign in / Sign up

Export Citation Format

Share Document