chinese room argument
Recently Published Documents


TOTAL DOCUMENTS

23
(FIVE YEARS 1)

H-INDEX

2
(FIVE YEARS 0)

2019 ◽  
pp. 254-263
Author(s):  
Alan J. McComas

This chapter considers the question of whether or not nonliving systems can acquire consciousness. It explores contemporary advances in technology, particularly in the field of artificial intelligence. The chapter also considers whether or not consciousness can be performed if inorganic matter replaced the components with which organisms experience consciousness. These and similar questions on nonhuman intelligence and consciousness are fleshed out with scenarios and thought experiments proposed throughout the 20th century, such as John Searle’s Chinese room argument and the archangel paradigm proposed by C. D. Broad. The chapter concludes with reflections on the human being’s inability to truly experience consciousness in the same way as nonhumans.


Author(s):  
Joshua Rust

John Rogers Searle (born July 31, 1932) is the Slusser Professor of Philosophy at the University of California, Berkeley. This analytic philosopher has made major contributions to the fields of the philosophy of mind, the philosophy of language, and social ontology. He is best known for his Chinese room argument, which aims to demonstrate that the formally described systems of computer functionalism cannot give rise to intentional understanding. Searle’s early work focused on the philosophy of language, where, in Speech Acts (1969), he explores the hypothesis that speaking a language is a rule-governed form of behavior. Just as one must follow certain rules in order to be considered to be playing chess, rules determine whether a speaker is making a promise, giving a command, asking a question, making a statement, and so forth. The kind of speech act that an utterance is depends on, among other conditions, its propositional content and illocutionary force. The content depicts the world as being a certain way, and the force specifies what a speaker is trying to do with that content. For example, for an utterance to qualify as a promise a speaker must describe a future act (content) and intend that the utterance place him or herself under an obligation to do that act (force). In Intentionality (1983), Searle argues that the structure of language not only mirrors but is derivative of the structure of intentional thought, so that core elements of his analysis of speech acts can be used as the basis for a theory of intentionality. Just as we can only promise by bringing certain propositional contents under a certain illocutionary force, intentional states such as belief, desire, fear, and joy can only be about the world in virtue of a representative content and a psychological mode. A theory of intentionality does not explain how intentionality is possible, given the basic facts of the world as identified by the natural sciences. Much of Searle’s work in the philosophy of mind, as found in Minds, Brains, and Science (1984) and The Rediscovery of the Mind (1992), is dedicated to the question of how mental facts, including but not limited to intentional facts, can be reconciled with basic, natural facts. Searle’s Chinese room argument is formulated in the service of rejecting computer functionalism, a prominent attempt at such reconciliation. Searle’s positive view, which he describes as "biological naturalism," is that mental facts are both caused by and features of underlying neurophysiological processes. In Speech Acts (1969), Searle claims that using language is akin to playing chess, in that both activities are made possible by participants following what he describes as "constitutive rules," rules that must be followed in order for someone to be considered to be undertaking those activities. Other institutional facts, such as money or the U.S. presidency, are also created and maintained in virtue of our following certain constitutive rules. For example, someone can only count as a U.S. president if that person is, among other conditions, a U.S. citizen who receives a majority of electoral votes. This thought is extended and explored in Searle’s two book-length contributions to the field of social ontology, The Construction of Social Reality (1995) and Making the Social World (2010). In addition to the philosophy of language and social ontology, Searle has made book-length contributions to the philosophy of action (Rationality in Action (2001)) and the philosophy of perception (Seeing Things as They Are: A Theory of Perception (2015)). He also famously engaged Jacques Derrida’s critique of J. L. Austin’s discussion of illocutionary acts ("Reiterating the Differences: A Reply to Derrida" (1977)). Searle has summarized his various positions in Mind, Language, and Society: Philosophy in the Real World (1998) and Mind: A Brief Introduction (2004).


Author(s):  
Robert Van Gulick

John Searle’s ‘Chinese room’ argument aims to refute ‘strong AI’ (artificial intelligence), the view that instantiating a computer program is sufficient for having contentful mental states. Imagine a program that produces conversationally appropriate Chinese responses to Chinese utterances. Suppose Searle, who understands no Chinese, sits in a room and is passed slips of paper bearing strings of shapes which, unbeknown to him, are Chinese sentences. Searle performs the formal manipulations of the program and passes back slips bearing conversationally appropriate Chinese responses. Searle seems to instantiate the program, but understands no Chinese. So, Searle concludes, strong AI is false.


2017 ◽  
Vol 10 (1) ◽  
pp. 38-49
Author(s):  
Corey Baron

This paper argues against John Searle in defense of the potential for computers to understand language (“Strong AI”) by showing that semantic meaning is itself a second-order system of rules that connects symbols and syntax with extralinguistic facts. Searle’s Chinese Room Argument is contested on theoretical and practical grounds by identifying two problems in the thought experiment, and evidence about “machine learning” is used to demonstrate that computers are already capable of learning to form true observation sentences in the same way humans do. Finally, sarcasm is used as an example to extend the argument to more complex uses of language


2017 ◽  
Vol 60 (1) ◽  
pp. 28-39
Author(s):  
Nenad Filipovic

The Chinese room argument is famous argument introduced by John Searle, in which Searle presented various problems with the claim that it is possible for the artificial intelligence to have understanding of a language in a way in which intelligent beings such as humans have that capacity. The argument was influential enough to, in decades following it, sparke numerous responses and critiques, along with a few alleged improvements to it from Searle. In this article, I will analyze one atypical critique of Searle?s argument, made by Mark Sprevak. Sprevak, unlike the other critics of the argument, agrees with Searle that understanding does not exist in Chinese room in any way, but he claims that Chinese room cannot execute every possible program, like Searle claims. Because of that, Searle cannot conclude the strong conclusion he wants from The Chinese room argument. In this article, I will analyze Searle?s argument, I will give a brief overview of typical responses to it, and I will analyze Sprevak?s response. In the last section, I will present argument that shows that Sprevak, if he wants to keep his conclusions, must either give up one part of his response, or accept one of the typical responses to Searle?s argument, thus making his own response dependent on the response from others.


Scholarpedia ◽  
2009 ◽  
Vol 4 (8) ◽  
pp. 3100 ◽  
Author(s):  
John Searle

Sign in / Sign up

Export Citation Format

Share Document