Thelanguage of thought hypothesis (LOTH),[1] sometimes known asthought ordered mental expression (TOME),[2] is a view inlinguistics,philosophy of mind andcognitive science, put forward by American philosopherJerry Fodor. It describes the nature of thought as possessing "language-like" or compositional structure (sometimes known asmentalese). On this view, simpleconcepts combine in systematic ways (akin to the rules ofgrammar in language) to buildthoughts. In its most basic form, the theory states that thought, like language, hassyntax.
Using empirical evidence drawn from linguistics and cognitive science to describemental representation from a philosophical vantage-point, the hypothesis states that thinking takes place in a language of thought (LOT):cognition and cognitive processes are only 'remotely plausible' when expressed as a system of representations that is "tokened" by a linguistic orsemantic structure and operated upon by means of acombinatorial syntax.[1] Linguistic tokens used in mental language describe elementary concepts which are operated upon bylogical rules establishingcausal connections to allow for complex thought. Syntax as well as semantics have a causal effect on the properties of this system of mental representations.
These mental representations are not present in thebrain in the same way assymbols are present on paper; rather, the LOT is supposed to exist at the cognitive level, the level of thoughts and concepts. The LOTH has wide-ranging significance for a number of domains incognitive science. It relies on a version of functionalist materialism, which holds that mental representations are actualized and modified by the individual holding the propositional attitude, and it challengeseliminative materialism andconnectionism. It implies a stronglyrationalist model of cognition in which many of the fundamentals of cognition are innate.[3][4][5]
The hypothesis applies to thoughts that have propositional content, and is not meant to describe everything that goes on in the mind. It appeals to the representational theory of thought to explain what those tokens actually are and how they behave. There must be a mental representation that stands in some unique relationship with the subject of the representation and has specific content. Complex thoughts get their semantic content from the content of the basic thoughts and the relations that they hold to each other. Thoughts can only relate to each other in ways that do not violate the syntax of thought. The syntax by means of which these two sub-parts are combined can be expressed infirst-order predicate calculus.
The thought "John is tall" is clearly composed of two sub-parts, the concept of John and the concept of tallness, combined in a manner that may be expressed in first-order predicate calculus as a predicate 'T' ("is tall") that holds of the entity 'j' (John). A fully articulated proposal for what a LOT would have to take into account greater complexities such asquantification andpropositional attitudes (the various attitudes people can have towards statements; for example I mightbelieve orsee or merelysuspect that John is tall).
The language of thought hypothesis has been both controversial and groundbreaking. Some philosophers reject the LOTH, arguing that our public languageis our mental language—a person who speaks Englishthinks in English. But others contend that complex thought is present even in those who do not possess a public language (e.g. babies,aphasics, and even higher-order primates), and therefore some form of mentalese must be innate.[citation needed]
The notion that mental states are causally efficacious diverges frombehaviorists likeGilbert Ryle, who held that there is no break between mental state and behavior (as cause and effect). Rather, Ryle proposed that people act in some way because they are in a disposition to act in that way, that these causal mental states are representational. An objection to this point comes fromJohn Searle in the form ofbiological naturalism, a non-representational theory of mind that accepts the causal efficacy of mental states. Searle divides intentional states into low-level brain activity and high-level mental activity. The lower-level, nonrepresentational neurophysiological processes have causal power in intention and behavior rather than some higher-level mental representation.[citation needed]
Tim Crane, in his bookThe Mechanical Mind,[6] states that, while he agrees with Fodor, his reason is very different. A logical objection challenges LOTH’s explanation of how sentences in natural languages get their meaning. That is the view that “Snow is white” is TRUE if and only ifP is TRUE in the LOT, whereP means the same thing in LOT as “Snow is white” means in the natural language. Any symbol manipulation is in need of some way of deriving what those symbols mean.[6] If the meaning of sentences is explained regarding sentences in the LOT, then the meaning of sentences in LOT must get their meaning from somewhere else. Thus there seems to be an infinite regress of sentences getting their meaning. Sentences in natural languages get their meaning from their users (speakers, writers).[6] Therefore, sentences in mentalese must get their meaning from the way in which they are used by thinkers and so onad infinitum. This regress is often called thehomunculus regress.[6]
Daniel Dennett accepts that homunculi may be explained by other homunculi and denies that this would yield an infinite regress of homunculi. Each explanatory homunculus is “stupider” or more basic than the homunculus it explains, but this regress is not infinite but bottoms out at a basic level that is so simple that it does not need interpretation.[6] John Searle points out that it still follows that the bottom-level homunculi are manipulating some sorts of symbols.
LOTH implies that the mind has sometacit knowledge of the logicalrules of inference and the linguistic rules ofsyntax (sentence structure) andsemantics (concept or word meaning).[6] If LOTH cannot show that the mind knows that it is following the particular set of rules in question, then the mind is not computational because it is not governed by computational rules.[3][6] Also, the apparent incompleteness of this set of rules in explaining behavior is pointed out. Many conscious beings behave in ways that are contrary to the rules of logic. Yet this irrational behavior is not accounted for by any rules, showing that there is at least some behavior that does not act by this set of rules.[6]
Another objection within representational theory of mind has to do with the relationship between propositional attitudes and representation. Dennett points out that a chess program can have the attitude of “wanting to get its queen out early,” without having representation or rule that explicitly states this. A multiplication program on a computer computes in the computer language of 1’s and 0’s, yielding representations that do not correspond with any propositional attitude.[3]
Susan Schneider has recently developed a version of LOT that departs from Fodor's approach in numerous ways. In her book, The Language of Thought: a New Philosophical Direction, Schneider argues that Fodor's pessimism about the success of cognitive science is misguided, and she outlines an approach to LOT that integrates LOT with neuroscience. She also stresses that LOT is not wedded to the extreme view that all concepts are innate. She fashions a new theory of mental symbols, and a related two-tiered theory of concepts, in which a concept's nature is determined by its LOT symbol type and its meaning.[4]
Connectionism is an approach toartificial intelligence that often accepts a lot of the same theoretical framework that LOTH accepts, namely that mental states are computational and causally efficacious and very often that they are representational. However, connectionism stresses the possibility of thinking machines, most often realized asartificial neural networks, an inter-connectional set of nodes, and describes mental states as able to create memory by modifying the strength of these connections over time. Some popular types of neural networks are interpretations of units, and learning algorithm. "Units" can be interpreted as neurons or groups of neurons. A learning algorithm is such that, over time, a change in connection weight is possible, allowing networks to modify their connections. Connectionist neural networks are able to change over time via their activation. An activation is a numerical value that represents any aspect of a unit that a neural network has at any time. Activation spreading is the spreading or taking over of other over time of the activation to all other units connected to the activated unit.
Since connectionist models can change over time, supporters of connectionism claim that it can solve the problems that LOTH brings to classical AI. These problems are those that show that machines with a LOT syntactical framework very often are much better at solving problems and storing data than human minds, yet much worse at things that the human mind is quite adept at such as recognizing facial expressions and objects in photographs and understanding nuanced gestures.[6] Fodor defends LOTH by arguing that a connectionist model is just some realization or implementation of the classicalcomputational theory of mind and therein necessarily employs a symbol-manipulating LOT.
Fodor andZenon Pylyshyn use the notion ofcognitive architecture in their defense. Cognitive architecture is the set of basic functions of an organism with representational input and output. They argue that it is a law of nature that cognitive capacities are productive, systematic and inferentially coherent—they have the ability to produce and understand sentences of a certain structure if they can understand one sentence of that structure.[7] A cognitive model must have a cognitive architecture that explains these laws and properties in some way that is compatible with the scientific method. Fodor and Pylyshyn say that cognitive architecture can only explain the property of systematicity by appealing to a system of representations and that connectionism either employs a cognitive architecture of representations or else does not. If it does, then connectionism uses LOT. If it does not then it is empirically false.[3]
Connectionists have responded to Fodor and Pylyshyn by denying that connectionism uses LOT, by denying that cognition is essentially a function that uses representational input and output or denying that systematicity is a law of nature that rests on representation.[citation needed] Some connectionists have developed implementational connectionist models that can generalize in a symbolic fashion by incorporating variables.[8]
This sectionneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources in this section. Unsourced material may be challenged and removed.(August 2011) (Learn how and when to remove this message) |
Since LOTH came to be it has been empirically tested. Not all experiments have confirmed the hypothesis;