Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Symbol Grounding Problem

View through CrossRef
The topic of representation acquisition, manipulation and use has been a major trend in Artificial Intelligence since its beginning and persists as an important matter in current research. Particularly, due to initial focus on development of symbolic systems, this topic is usually related to research in symbol grounding by artificial intelligent systems. Symbolic systems, as proposed by Newell & Simon (1976), are characterized as a highlevel cognition system in which symbols are seen as “[lying] at the root of intelligent action” (Newell and Simon, 1976, p.83). Moreover, they stated the Physical Symbol Systems Hypothesis (PSSH), making the strong claim that “a physical symbol system has the necessary and sufficient means for general intelligent action” (p.87). This hypothesis, therefore, sets equivalence between symbol systems and intelligent action, in such a way that every intelligent action would be originated in a symbol system and every symbol system is capable of intelligent action. The symbol system described by Newell and Simon (1976) is seen as a computer program capable of manipulating entities called symbols, ‘physical patterns’ combined in expressions, which can be created, modified or destroyed by syntactic processes. Two main capabilities of symbol systems were said to provide the system with the properties of closure and completeness, and so the system itself could be built upon symbols alone (Newell & Simon, 1976). These capabilities were designation – expressions designate objects – and interpretation – expressions could be processed by the system. The question was, and much of the criticism about symbol systems came from it, how these systems, built upon and manipulating just symbols, could designate something outside its domain. Symbol systems lack ‘intentionality’, stated John Searle (1980), in an important essay in which he described a widely known mental experiment (Gedankenexperiment), the ‘Chinese Room Argument’. In this experiment, Searle places himself in a room where he is given correlation rules that permits him to determine answers in Chinese to question also in Chinese given to him, although Searle as the interpreter knows no Chinese. To an outside observer (who understands Chinese), the man in this room understands Chinese quite well, even though he is actually manipulating non-interpreted symbols using formal rules. For an outside observer the symbols in the questions and answers do represent something, but for the man in the room the symbols lack intentionality. The man in the room acts like a symbol system, which relies only in symbolic structures manipulation by formal rules. For such systems, the manipulated tokens are not about anything, and so they cannot even be regarded as representations. The only intentionality that can be attributed to these symbols belongs to who ever uses the system, sending inputs that represent something to them and interpreting the output that comes out of the system. (Searle, 1980) Therefore, intentionality is the important feature missing in symbol systems. The concept of intentionality is of aboutness, a “feature of certain mental states by which they are directed at or about objects and states of affairs in the world” (Searle, 1980), as a thought being about a certain place.1 Searle (1980) points out that a ‘program’ itself can not achieve intentionality, because programs involve formal relations and intentionality depends on causal relations. Along these lines, Searle leaves a possibility to overcome the limitations of mere programs: ‘machines’ – physical systems causally connected to the world and having ‘causal internal powers’ – could reproduce the necessary causality, an approach in the same direction of situated and embodied cognitive science and robotics. It is important to notice that these ‘machines’ should not be just robots controlled by a symbol system as described before. If the input does not come from a keyboard and output goes to a monitor, but rather came in from a video camera and then out to motors, it would not make a difference since the symbol system is not aware of this change. And still in this case, the robot would not have intentional states (Searle 1980). Symbol systems should not depend on formal rules only, if symbols are to represent something to the system. This issue brought in another question, how symbols could be connected to what they represent, or, as stated by Harnad (1990) defining the Symbol Grounding Problem: “How can the semantic interpretation of a formal symbol system be made intrinsic to the system, rather than just parasitic on the meanings in our heads? How can the meanings of the meaningless symbol tokens, manipulated solely on the basis of their (arbitrary) shapes, be grounded in anything but other meaningless symbols?” The Symbol Grounding Problem, therefore, reinforces two important matters. First that symbols do not represent anything to a system, at least not what they were said to ‘designate’. Only someone operating the system could recognize those symbols as referring to entities outside the system. Second, the symbol system cannot hold its closure in relating symbols only with other symbols; something else should be necessary to establish a connection between symbols and what they represent. An analogy made by Harnad (1990) is with someone who knows no Chinese but tries to learn Chinese from a Chinese/Chinese dictionary. Since terms are defined by using other terms and none of them is known before, the person is kept in a ‘dictionary-goround’ without ever understanding those symbols. The great challenge for Artificial Intelligence researchers then is to connect symbols to what they represent, and also to identify the consequences that the implementation of such connection would make to a symbol system, e.g. much of the descriptions of symbols by means of other symbols would be unnecessary when descriptions through grounding are available. It is important to notice that the grounding process is not just about giving sensors to an artificial system so it would be able to ‘see’ the world, since it ‘trivializes’ the symbol grounding problem and ignores the important issue about how the connection between symbols and objects are established (Harnad, 1990).
Title: Symbol Grounding Problem
Description:
The topic of representation acquisition, manipulation and use has been a major trend in Artificial Intelligence since its beginning and persists as an important matter in current research.
Particularly, due to initial focus on development of symbolic systems, this topic is usually related to research in symbol grounding by artificial intelligent systems.
Symbolic systems, as proposed by Newell & Simon (1976), are characterized as a highlevel cognition system in which symbols are seen as “[lying] at the root of intelligent action” (Newell and Simon, 1976, p.
83).
Moreover, they stated the Physical Symbol Systems Hypothesis (PSSH), making the strong claim that “a physical symbol system has the necessary and sufficient means for general intelligent action” (p.
87).
This hypothesis, therefore, sets equivalence between symbol systems and intelligent action, in such a way that every intelligent action would be originated in a symbol system and every symbol system is capable of intelligent action.
The symbol system described by Newell and Simon (1976) is seen as a computer program capable of manipulating entities called symbols, ‘physical patterns’ combined in expressions, which can be created, modified or destroyed by syntactic processes.
Two main capabilities of symbol systems were said to provide the system with the properties of closure and completeness, and so the system itself could be built upon symbols alone (Newell & Simon, 1976).
These capabilities were designation – expressions designate objects – and interpretation – expressions could be processed by the system.
The question was, and much of the criticism about symbol systems came from it, how these systems, built upon and manipulating just symbols, could designate something outside its domain.
Symbol systems lack ‘intentionality’, stated John Searle (1980), in an important essay in which he described a widely known mental experiment (Gedankenexperiment), the ‘Chinese Room Argument’.
In this experiment, Searle places himself in a room where he is given correlation rules that permits him to determine answers in Chinese to question also in Chinese given to him, although Searle as the interpreter knows no Chinese.
To an outside observer (who understands Chinese), the man in this room understands Chinese quite well, even though he is actually manipulating non-interpreted symbols using formal rules.
For an outside observer the symbols in the questions and answers do represent something, but for the man in the room the symbols lack intentionality.
The man in the room acts like a symbol system, which relies only in symbolic structures manipulation by formal rules.
For such systems, the manipulated tokens are not about anything, and so they cannot even be regarded as representations.
The only intentionality that can be attributed to these symbols belongs to who ever uses the system, sending inputs that represent something to them and interpreting the output that comes out of the system.
(Searle, 1980) Therefore, intentionality is the important feature missing in symbol systems.
The concept of intentionality is of aboutness, a “feature of certain mental states by which they are directed at or about objects and states of affairs in the world” (Searle, 1980), as a thought being about a certain place.
1 Searle (1980) points out that a ‘program’ itself can not achieve intentionality, because programs involve formal relations and intentionality depends on causal relations.
Along these lines, Searle leaves a possibility to overcome the limitations of mere programs: ‘machines’ – physical systems causally connected to the world and having ‘causal internal powers’ – could reproduce the necessary causality, an approach in the same direction of situated and embodied cognitive science and robotics.
It is important to notice that these ‘machines’ should not be just robots controlled by a symbol system as described before.
If the input does not come from a keyboard and output goes to a monitor, but rather came in from a video camera and then out to motors, it would not make a difference since the symbol system is not aware of this change.
And still in this case, the robot would not have intentional states (Searle 1980).
Symbol systems should not depend on formal rules only, if symbols are to represent something to the system.
This issue brought in another question, how symbols could be connected to what they represent, or, as stated by Harnad (1990) defining the Symbol Grounding Problem: “How can the semantic interpretation of a formal symbol system be made intrinsic to the system, rather than just parasitic on the meanings in our heads? How can the meanings of the meaningless symbol tokens, manipulated solely on the basis of their (arbitrary) shapes, be grounded in anything but other meaningless symbols?” The Symbol Grounding Problem, therefore, reinforces two important matters.
First that symbols do not represent anything to a system, at least not what they were said to ‘designate’.
Only someone operating the system could recognize those symbols as referring to entities outside the system.
Second, the symbol system cannot hold its closure in relating symbols only with other symbols; something else should be necessary to establish a connection between symbols and what they represent.
An analogy made by Harnad (1990) is with someone who knows no Chinese but tries to learn Chinese from a Chinese/Chinese dictionary.
Since terms are defined by using other terms and none of them is known before, the person is kept in a ‘dictionary-goround’ without ever understanding those symbols.
The great challenge for Artificial Intelligence researchers then is to connect symbols to what they represent, and also to identify the consequences that the implementation of such connection would make to a symbol system, e.
g.
much of the descriptions of symbols by means of other symbols would be unnecessary when descriptions through grounding are available.
It is important to notice that the grounding process is not just about giving sensors to an artificial system so it would be able to ‘see’ the world, since it ‘trivializes’ the symbol grounding problem and ignores the important issue about how the connection between symbols and objects are established (Harnad, 1990).

Related Results

PENGARUH KELEMBABAN TANAH TERHADAP TAHANAN PENTANAHAN STUDI KASUS PADA GARDU INDUK KEMAYORAN 150 kV
PENGARUH KELEMBABAN TANAH TERHADAP TAHANAN PENTANAHAN STUDI KASUS PADA GARDU INDUK KEMAYORAN 150 kV
Abstract The value of grounding resistance at the substation should be 0 Ω or less than 1 Ω. The value of grounding resistance is influenced by the resistivity and the ground...
MAKNA SIMBOL DAN ORNAMEN PADA BANGUNAN VIHARA GUNUNG TIMUR
MAKNA SIMBOL DAN ORNAMEN PADA BANGUNAN VIHARA GUNUNG TIMUR
The title of this research was ‘The Meaning of Symbols and Ornaments in the Building of Vihara Gunung Timur’. The research problems were 1) what types of symbol and ornament are fo...
Grounding and Calving Cycle of Mertz Ice Tongue Revealed by Shallow Mertz Bank
Grounding and Calving Cycle of Mertz Ice Tongue Revealed by Shallow Mertz Bank
Abstract. A recent study, using remote sensing, provided some evidence that a seafloor shoal influenced the 2010 calving event of the Mertz Ice Tongue (MIT), by partially grounding...
Minimizing Grounding Resistance of Cathodic Protection Anode Bed with Finite Element Method
Minimizing Grounding Resistance of Cathodic Protection Anode Bed with Finite Element Method
Abstract Cathodic protection (CP) anode bed grounding resistance calculating formulae, given by CP standards, are reviewed firstly. Underground assumptions and limit...
Representation of basal melting at the grounding line in ice flow models
Representation of basal melting at the grounding line in ice flow models
Abstract. While a lot of attention has been given to the numerical implementation of grounding lines and basal friction in the grounding zone, little has been done about the impact...
THE OPTIMISTIC SYMBOLISMS IN ERNEST HEMINGWAY’S THE OLD MAN AND THE SEA
THE OPTIMISTIC SYMBOLISMS IN ERNEST HEMINGWAY’S THE OLD MAN AND THE SEA
This study aims at explaining the symbols used in the novel and describing the values of optimistic reflected in the symbols used in Hemingway’s The Old man and The Sea. This resea...
Grounding and calving cycle of Mertz Ice Tongue revealed by shallow Mertz Bank
Grounding and calving cycle of Mertz Ice Tongue revealed by shallow Mertz Bank
Abstract. A recent study, using remote sensing, provided evidence that a seafloor shoal influenced the 2010 calving event of the Mertz Ice Tongue (MIT), by partially grounding the ...

Back to Top