Chinese Room Argument in Artificial Intelligence
When we ask, ‘Is artificial intelligence (AI) possible?’ We really ask ‘Can we create consciousness in computers’ ?
The Chinese room argument holds that a program cannot give a computer a “mind”, “understanding” or “consciousness”” regardless of how intelligently or human-like the program may make the computer behave/ [Source Wiki]
In 1980, John Searle argued that Turing Test could not be used to determine “whether or not a machine is considered as intelligent like humans”. He argued that any machine like ELIZA and PARRY could easily pass Turing Test simply by manipulating symbols of which they had no understanding. Without understanding, they could not be described as “thinking” in the same sense people do.
John imagines himself (instead of machine) as non-Chinese person sitting inside the room isolated from another Chinese person who is outside the room tries to communicate. He is provided a list of Chinese characters and an instruction book explaining in detail the rules according to which strings (sequences) of characters may be formed, but without giving the meaning of the characters. That means he has a book with an English version of the computer program, along with sufficient paper, pencils, erasers, and filing cabinets.
In this thought experiment, a person in the “Chinese room” is passed questions from outside the room, and consults a library of books to formulate an answer
Now he receive all the messages posted through a slot in the door written in Chinese language. He will process all the symbols according to program instructions and produces the chinese characters as output like:
- If he finds Chinese symbol like ♍, he returns symbol ♓
- If he finds Chinese symbol like ԪԘ, he returns symbol ԈԂ
Actually instruction book contains so many rules that contains input symbols and their respective output symbol. He just need to locate the input Chinese symbol and return the corresponding Chinese symbol as a output.
Now, the argument goes on, a computer(machine), is just like this man, in that it does nothing more than follow the rules given in an instruction book (the program). It does not understand the meaning of the questions given to it nor its own answers, and thus cannot be said to be thinking. The fact is that inside person has no understanding of Chinese language but still he manage to communicate with outside person in Chinese language perfectly.
Compare the John with machine in Turing Test, the machine may have huge collection of database containing questions and answers. When a interrogator ask the question, the machine is simply locating the question in the database and returning the corresponding answer to the interrogator. The whole scenario would seems like that human is returning the answer unlike machine.
Hence the machine in configuration has no understanding of those questions and answers, without “understanding” (or “intentionality“), we cannot describe what the machine is doing as “thinking” and, since it does not think, it does not have a “mind” in anything like the normal sense of the word. Therefore we can’t consider machine as intelligent.
This article is contributed by Shubham Bansal. If you like GeeksforGeeks and would like to contribute, you can also write an article using write.geeksforgeeks.org or mail your article to contribute@geeksforgeeksorg. See your article appearing on the GeeksforGeeks main page and help other Geeks.