Logic and Language
Load the menuLoad the menu


Copyright   James R Meyer    2012 - 2020 https://www.jamesrmeyer.com

Note: Full functionality of this website requires JavaScript to be enabled in your browser.
 

The Chinese Room

Page last updated 13 June 2021

 

In 1980, the philosopher John Searle published a paper Minds, Brains and Programs that claimed to show that artificial intelligence machines could never have ‘understanding’, regardless of their reasoning abilities. The fundamental idea of John Searle’s argument (commonly called the Chinese Room argument), taken down to its bare bones, is as follows:

Consider a thing with the following properties:

Chinese art
  1. The thing can take, as an input, sentences of a given language.
  2. The thing includes the ability to process the input and generate output sentences of that same language according to a set of instructions.
  3. The thing includes as an integral part of itself this set of instructions. The instructions determine what the output will be for any valid input. The instructions are such that, for valid inputs, outputs are generated that would be an appropriate response for a human fluent in that language.
  4. The thing always follows the instructions.

From this Searle claims:

Premise: the processing part of the thing does not require any understanding of the language in order to process the input and generate an output according to the instructions,

therefore

Conclusion: the thing does not understand the language.

 

The absurdity of the argument is obvious. There is no logical basis, given the premise, for inferring the conclusion. There is no reasoned argument at all.

 

It is quite astonishing that there has been ongoing controversy over Searle’s Chinese room argument for so long when a logical analysis of the argument readily demonstrates the absurdity of the argument.

 

Perhaps the reason is that the argument which Searle presents dresses up the above fundamental argument by adding in quite extraneous details which do not affect the basic premises. In Searle’s account, the thing is a closed room Chinese symbols with no windows, the inputs and outputs are slits in the wall through which paper can be pushed, the language is Chinese, Chinese symbolsthe set of instructions are written on paper, and the processing part of the thing is a human who does not understand Chinese. If the instructions are good enough, the responses of the room will be indistinguishable from the responses of a fluent Chinese human.

 

Searle concludes that, although the responses are indistinguishable from the responses of a fluent Chinese human, since the human in the room does not understand Chinese, then the entire room does not understand Chinese.

 

But when the extra baggage is removed from the argument, we see that Searle’s conclusion is simply the conclusion that he wants, reached by an appeal to intuition. Searle’s Chinese Room does not provide any logical basis for his conclusion, but serves as a smoke screen that has obscured the fact that Searle has no logical basis whatsoever for his conclusion.

 

Searle admits that there might be objections to his arguments on various grounds, and gives counter-arguments to those objections. None of the counter-arguments that he refers to address the issue of the lack of any logical inference of Searle’s conclusion from his premise.

 

It is unfortunate that responses to Searle seem to have concentrated on every other aspect of his argument other than his failure to provide a logical reason for drawing his conclusion from the premises given. Such responses have unfortunately created the impression that there might be some substance in his argument, whereas a logical analysis shows his argument to be hopelessly subjective and without any logical foundation.

 

The failure of Searle and others to perceive that the argument fails to logically derive its conclusion from the premise appears to result from the lack of definition of ‘understanding a language’ or any measurement of such. An intuitive notion of ‘understanding’ is applied instead. Nowhere does Searle make any attempt to move towards an objective definition of what he intends ‘to understand a language’ to mean, nor does he give any consideration to the question of formulating an objective method of measurement of ‘understanding’ of a language. Instead Searle boldly asserts that he knows what ‘to understand a language’ means, but refuses to define it, with the result no-one can be sure what he is actually talking about. Searle asserts that his intuition and subjective judgment are to be the criteria by which it is to be known that no machine could ever be made that could understand a language, where ‘to understand a language’ means only what Searle wants it to mean in any given context, and is not be subjected to any objective measurement.

 

Understanding

While Searle’s Chinese Room scenario cannot be used to prove that no machine can ever ‘understand’ a human language, it nevertheless raises interesting issues regarding what we mean by ‘understanding’. Searle gives no objective measure of understanding of Chinese that we can apply to the processor and the entire thing. Searle blithely states that the entire thing understands nothing of Chinese, but does not consider how the entire thing might respond to an objective test of understanding of Chinese.

 

What Searle calls the ‘other minds reply’ is perhaps the closest he comes to considering the definition of understanding. Searle states that an objection might be:

“Searle has not given any information as to how one might determine whether another mind understands Chinese, that one might expect that such determination would have to be by the behaviour of that mind, that Searle has not demonstrated that a machine can never pass such behavioural tests as well as a human.”

Searle dismisses this in a few sentences, without addressing the issue of measurement of understanding, and confuses the issue by referring to ‘cognitive states’ rather than ‘understanding’ (continually changing the terms of reference is a common method of deflecting an inconvenient objection to an argument).

 

Searle states that (here I have reworded his arguments since his original arguments are so abstruse and poorly worded)

  1. One needs to consider not how I know that people/machines have understanding, but rather what it is that I am attributing to them when I attribute understanding to them.

Here Searle deflects the question of an objective measurement of understanding, and dismisses it as unimportant. He considers his subjective judgment to be sufficient for determining what understanding is.

  1. When one states that a human has ‘understanding’, the attributes that underpin ‘understanding’ cannot be merely computational processes and their output, because the computational processes and their output can exist without understanding.

Here Searle argues that since computational processes and their output can exist without understanding, that no combination of computational processes and their outputs cannot ever be considered to have ‘understanding’. This argument is absurd. One might as well say that since parts of the brain can exist without having understanding, then the brain as a complete entity cannot have understanding. Or that since muscles cannot walk, then legs which use muscles cannot walk.

  1. In ‘understanding’ one must assume that there is a reality and knowability of the mental state in the same way that in physical sciences one has to assume the reality and knowability of physical objects.

Here Searle simply argues that one must assume that there is a ‘mental state’ that is real. But that ‘mental state’ is not itself a physical object, rather it is the property of a physical object, so the ‘mental state’ is dependent for its existence on the physical attributes of the physical object. Searle does not explain how this might refute the objection, unless he is assuming that no machine, regardless of its physical attributes, can have a ‘mental state’. But that is simply introducing another ill-defined term, ‘mental state’, which adds nothing to Searle’s argument and serves only to confuse. Other attempts by Searle to deal with objections to his argument show similar flawed logic and similar appeals to intuition.

Note: A Bruce R Bain who manages an average of over 6 posts per day over the last 5 years has posted on Quora Why does the Gödel critic James Meyer object to John Searle’s Chinese room argument? Is his refutation of Searle convincing? with the complaint that I assume that a ‘mental state’ is the property of a physical object. The post is laughable, complaining that I do not refer to any “peer reviewed science publication showing that mental states are the property of a physical object.” Of course I don’t - it was Searle, not me, who introduced the term ‘mental state’, and Searle himself actually states in his paper: (Footnote: If a “mental state” is not something that is the property of a physical object, then the only options are that it is either a non-physical “real” entity, or the property of some non-physical “real” entity. But if Searle’s intention was to assume some sort of non-physical “real” entity, then that is all he is left with - an assumption of the existence of non-physical “real” entities - and the Chinese Room argument, which is an argument that is based only on physical entities and the physical actions that those entities perform, then becomes completely irrelevant. )

…mental states and events are literally a product of the operation of the brain…

thus confirming that he views a “mental state” as the property of a physical thing.

 

Another interesting viewpoint on the Chinese Room argument can also be seen at Conscious Entities - Against the Chinese Room.

 

Other Posts

section divider

Footnotes:

section divider

 

 

As site owner I reserve the right to keep my comments sections as I deem appropriate. I do not use that right to unfairly censor valid criticism. My reasons for deleting or editing comments do not include deleting a comment because it disagrees with what is on my website. Reasons for exclusion include:
Frivolous, irrelevant comments.
Comments devoid of logical basis.
Derogatory comments.
Long-winded comments.
Comments with excessive number of different points.
Questions about matters that do not relate to the page they post on. Such posts are not comments.
Comments with a substantial amount of mathematical terms not properly formatted will not be published unless a file (such as doc, tex, pdf) is simultaneously emailed to me, and where the mathematical terms are correctly formatted.


Reasons for deleting comments of certain users:
Bulk posting of comments in a short space of time, often on several different pages, and which are not simply part of an ongoing discussion. Multiple anonymous usernames for one person.
Users, who, when shown their point is wrong, immediately claim that they just wrote it wrong and rewrite it again - still erroneously, or else attack something else on my site - erroneously. After the first few instances, further posts are deleted.
Users who make persistent erroneous attacks in a scatter-gun attempt to try to find some error in what I write on this site. After the first few instances, further posts are deleted.


Difficulties in understanding the site content are usually best addressed by contacting me by e-mail.

 

Note: a password enables editing of comments, an email enables notification of replies

HashOver logoBased on HashOver Comment System by Jacob BarkdullHashOver logo

Copyright   James R Meyer   2012 - 2020
https://www.jamesrmeyer.com