Abstracts for SPP Symposium On Searle's Chinese Room
and
Workshop on Symbol Grounding
Saturday, June 9, University of Maryland, College Park
(Full Program Follows in Next Posting)
For information: andrewsj at vassar.bitnet
----------------------------------------------------------------------
(1) SYNTAX, SEMANTICS AND SYSTEMS
John McCarthy <JMC at SAIL.Stanford.EDU>
Artificial Intelligence
Stanford University
John Searle begins his (1990) "Consciousness, Explanatory Inversion and
Cognitive Science" with:
"Ten years ago in this journal I published an
article (Searle, 1980a and 1980b) criticising what I
call Strong AI, the view that for a system to have
mental states it is sufficient for the system to
implement the right sort of program with right inputs
and outputs. Strong AI is rather easy to refute and
the basic argument can be summarized in one sentence:
`a system, me for example, could implement a program for
understanding Chinese, for example, without
understanding any Chinese at all.' This idea, when
developed, became known as the Chinese Room Argument."
The Chinese Room Argument can be refuted in one sentence:
"Searle confuses the mental qualities of one computational process,
himself for example, with those of another process that the first
process might be interpreting, a process that understands Chinese, for
example."
That accomplished, the lecture will discuss the ascription of mental
qualities to machines with special attention to the relation between
syntax and semantics, i.e. questions suggested by the Chinese Room
Argument. I will deal explicitly with Searle's four ``axioms'',
which, although they don't have a unique interpretation, suggest
various ideas worth discussing.
-----------------------------------------------------------
(2) SEARLE'S CHINESE ROOM
Pat Hayes <hayes at parc.xerox.com>
The basic flaw in Searle's argument is a widely accepted misunderstanding
about the nature of computers and computation: the idea that a computer
is a mechanical slave that obeys orders. This popular metaphor suggests
a major division between physical, causal hardware which acts, and
formal symbolic software, which gets read. This distinction runs
through much computing terminology, but one of the main conceptual
insights of computer science is that it is of little real scientific
importance. Computers running programs just aren't like the Chinese
room.
Software is a series of patterns which, when placed in the proper
places inside the machine, cause it to become a causally different
device. Computer hardware is by itself an incomplete specification of a
machine, which is completed - i.e. caused to quickly reshape its
electronic functionality - by having electrical patterns moved within
it. The hardware and the patterns together become a mechanism which
behaves in the way specified by the program.
This is not at all like the relationship between a reader obeying some
instructions or following some rules. Unless, that is, he has somehow
absorbed these instructions so completely that they have become part of
him, become one of his skills. The man in Searle's room who has done
this to his program now understands Chinese.
--------------------------------------------------------------------
(3)E MEASUREMENT PROBLEM IN PHSYICS AND BRAIN THEORY
H.H. Pattee
Department of Systems Science
TJ Watson School of Engineering
SUNY
Binghamton NY 13901
The measurement problem in physics is a special case of the
symbol-grounding problem of brain theory, which in turn is a special
case of the epistemological problem of relating the knower and the
known. In quantum theory the measurement problem is notoriously
obscure, because the results of measuring quantum events are
nonclassically observer-dependent and yet must be expressed only in
classical language. Even the measurement of classical events cannot be
completely described by physical laws, because measurement involves
intent, i.e., the what, where, and when of the measurements must be
determined by an "observer," not by the laws. It also makes no sense
to say that a measurement has occurred unless there is a "result."
At issue in physical theory are the necessary conditions for
"observer," and "result." Working physicists evade the philosophical
issues by using language confined to formal, operational symbol
systems that restrict measurements to numerical results and
prediction to computation. Formal symbol systems do allow unambiguous
predictions, but only at the cost of generating logical antinomies and
conceptual puzzles like "Schroedinger's Cat" and "Wigner's friend."
If one relaxes the syntactic precision of formal symbols, however, and
extends the concepts of "observer" and "result" to the subsymbolic,
dynamical level of measurement-control constraints in simpler
organisms, some of these puzzles are resolved. The problem I raise for
brain theory is that formal symbolic models of the brain may likewise
produce unambiguous results only at the cost of conceptual puzzles like
Searle's "Chinese Room."
------------------------------------------------------------------
(4) METAPHOR AND SYMBOL
David Powers <powers at informatik.uni-kl.de>
Robotics
University of Kaiserslautern FRG
The presentation is based on the monograph [Powe89] on
Machine Learning of Natural Language and Ontology,
updated with recent developments and indication of
consolidation of earlier work into a coherent approach.
Hypotheses concerning the role of contrast and
similarity, and the relationship of these mechanisms to linguistic
concepts of metaphor and paradigm, neural self-organization,
psycholinguistic paradoxes concerning negative information,
and consideration of language as part of the entire ontology
have lead to a series of experiments in machine learning of
aspects of language both individually or in combination. The work
makes use of a simulated robot world as well as textual input.
Significantly, similar results have been achieved with neural and
conventional techniques applied to the same task, with simulated
neurons being clearly associated with words and classes, and
with particular grammatical rules being associated with
particular synapses.
These results suggest three possible resolutions of the symbol
grounding problem: the symbol/non-symbol distinction
is not meaningful; neural networks can exhibit 'symbolic' behaviour
and structure; and, a sensory-motor environment can provide
grounding.
REFERENCES
[Powe89] David M. W. Powers, C. C. R. Turk,
Machine Learning of Natural Language, Springer Verlag,
London/Berlin, 1989.
-------------------------------------------------------------------
(5) CATEGORIZATION AS A PSYCHOPHYSICAL PROBLEM
Michel Treisman treisman at vax.oxford.ac.uk
Psychology Department
University of Oxford
If the same observer has to categorize the same stimulus on two
different occasions he may make different decisions each time. Why is
this? In some situations the observer will show an excessive tendency
to repeat previous responses to similar stimuli; this is sometimes
referred to as 'assimilation'. At other times he or she may avoid
previous responses: 'contrast'. Why should categorization be so
unreliable? Or what does this observation tell us about the process of
making a judgment? A psychophysical model will be outlined which
provides an explanation for these phenomena in terms of mechanisms
which tend to optimize uncertain judgments, and the relations between
different types of categorization, at different levels of complexity,
will be considered.
---------------------------------------------------------------------
(6) GROUNDING MENTAL SYMBOLS IN OBJECT IMAGES
Irving Biederman <PSYIRV at vx.acs.umn.edu> and John E. Hummel
Psychology Department
University of Minnesota
We describe a neural net (NN) implementation of a theory of real time
visual shape recognition that takes as input the edges corresponding to
the occlusional and orientation discontinuities in an image. As output
the model activates a unit that is selective for a specified
arrangement of simple volumes (or geons) and thus achieves a basic (or
entry) level classification according to Biederman's (1987)
Recognition-by-Components theory of object recognition. The output unit
can qualify as a symbol of the object in that it reflects the major
invariances of visual object recognition. The model solves four
fundamental problems in object recognition that likely confront all
attempts at visual basic-level symbol grounding:
1) Translational, size, and orientation invariance: The same output
unit(s), corresponding to the object, are activated no matter where the
image falls in the visual field, the size of the image, and the
orientation in depth (up to parts occlusion),
2) Appropriate grouping (or organization) of image elements into
appropriate parts,
3) A basis of determining invariant object centered relations (such as
TOP- OF or SIDE-CONNECTED), and
4) A basis for computing the similarity (or equivalence) of object
images.
These problems all required a solution to the "binding problem"--
determining what groups with what. In the present case, for example,
how are the various segments of the parts of an object grouped
according to their appropriate parts. Most NN models have employed
enumeration, assigning a unit to each attribute combination. Such
enumerative schemes are unsatisfactory in that they require a
prohibitively large number of units to represent even modest input
domains. Moreover, they do not express the equivalence of inputs. By
employing different units to represent the different locations of an
object, for example, the information that it is the same object in the
different locations is not represented. Our model achieves binding
through phase locking of the oscillatory activity of cells that are
tuned to oriented image edges. The phase locking (or synchrony) is
established by "fast enabling links" (FELs) between pairs of a)
collinear, b) coterminating, and c) parallel adjacent edge cells. These
units then activate invariant representations of geons and relations in
intermediate layers.
The model offers some perspective on what it is that makes a category
"basic." A category such as chair will encompass a number of
distinguishably different geon models that, perceptually, may be as
distant as members of different classes.
------------------------------------------------------
(7) INVARIANT STIMULUS FEATURES AND THE
CORTICAL REPRESENTATION OF VISUAL INFORMATION
J. Anthony Movshon tony at cortex.psych.nyu.edu
Center for Neural Science
New York University,
New York, NY 10003.
The responses of visual cortical neurons depend upon a number of
different features of the visual stimuli that fall within their
receptive fields. In most cases there is no difference in the nature
of the response produced by varying the stimulus along different
dimensions. For this reason, it is commonly recognized that the
firing of an individual neuron cannot be used unambiguously to infer
the character of a particular visual stimulus. Rather, it is necessary
to examine the distribution of activity across a population of
neurons. In considering how the multi-dimensional nature of visual
neural signals might most readily be disambiguated, it seems that
special significance might be attached to those stimulus dimensions
for which particular groups of neurons show an invariant selectivity.
An invariant selectivity is a selectivity for the value of a stimulus
along some dimension that is independent of the value of the stimulus
along other dimensions. For example, the selectivity of neurons in
the primary visual cortex (V1) for such stimulus variables as
orientation and spatial frequency is largely independent of the
precise stimulus conditions used to measure them. On the other hand,
their selectivity for the direction of motion of targets depends in a
complex way on the spatial and temporal composition of the target, and
is therefore not invariant.
In the visual cortex of the macaque monkey, many distinct visual
areas have been identified with electrophysiological and anatomical
techniques. A number of ``lower-order'' cortical areas seem to
contain neurons whose activity is primarily controlled by signals of
retinal origin - prominent among these are areas V2, V3, V4 and MT, as
well as the primary visual cortex, V1. The responses of neurons in all
these areas seem to depend on the same collection of visual stimulus
dimensions, including spatial location and size, contour orientation,
spatial frequency, chromatic composition, drift rate, direction of
motion, and binocular disparity. Neurons in different areas can have
more or less sensitivity to variations in one or another of these
parameters, so that in quantitative terms it may be argued that
signals from one area carry more information than signals from another
about particular stimulus features. It is largely on the basis of
quantitative arguments of this kind that a particular role for one or
another area in a particular aspect of visual processing has been
asserted - qualitative differences in the way that visual signals are
represented in different cortical areas have not received much
attention.
In this paper I will argue that a special significance is, in
fact, attached to the particular stimulus dimensions that are the
subjects of invariant representation within an area. For example,
despite the fact that neurons in V1 carry signals about the direction
and speed of motion of objects, the fact that their invariant
selectivity is for orientation, spatial and temporal frequency makes
it impossible for them to carry invariant information about speed and
direction. Neurons in MT, on the other hand, carry invariant
information about motion, at the expense of losing the invariant
representation of spatial and temporal parameters. Analogous
reorganization of signals about color, stereoscopic depth, and other
stimulus features may explain the existence of other representations
of the visual image in the visual cortex.
-----------------------------------------------------------
(8) COLOR AND COLOR CONSTANCY
Laurence T. Maloney ltm at xp.psych.nyu.edu
Department of Psychology
Center for Neural Science
New York University
The initial visual information that determines color appearance in
human vision depends as much on the lighting in a scene as on the
spectral properties of surfaces in the scene. A visual system that
bases color appearance on the properties of the surface, discounting
the contribution of the illuminant, is termed \fIcolor constant\fP.
I describe a class of algorithms designed to allow vision systems to
estimate information (analogous to color) about surface properties
despite changes in the illuminant. These linear model algorithms
include work by Brill, Buchsbaum, Maloney and Wandell, and others.
These algorithms share strong assumptions about the range of possible
illuminants and possible surface reflectances present in a scene. I
describe evidence suggesting that many common surfaces and illuminants
satisfy the constraints required by linear model algorithms.
Hilbert (1987, Chap. 7) discusses the consequences of this work
for philosophy.
Maloney, L. T., and Wandell, B. A., Color constancy: A computational
method for recovering surface spectral reflectance. Journal of the
Optical Society of America A, 1986, \fB3\fP, 29-33.
Maloney, L. T., Evaluation of linear models of surface spectral
reflectance with small numbers of parameters. Journal of the Optical
Society of America A, 1986, \fB3\fP, 1673-1683.
Hilbert, D. R., \fIColor and Color Perception; A Study in
Anthropocentric Realism.\fP (Stanford, CA: CSLI, 1987).
---------------------------------------------------
(9) PERCEPTUAL MEMORY CATEGORIZATION IN PRIMARY SENSORY CORTEX
Richard Granger granger at ICS.UCI.EDU
Center for the Neurobiology of Learning and Memory
University of California, Irvine
Recent results from neurobiological simulation work have led to a novel
hypothesis: that the physiological operation of a primary sensory
cortical area (olfactory (piriform) cortex) automatically organizes
learned perceptual cues into a hierarchical memory (Ambros-Ingerson,
Granger and Lynch, Science, 1990). In the simulation, repetitive
perceptual samples ("sniffs", or "glances") of learned cues traverse
the constructed hierarchy, such that initial samples yield relatively
coarse-grained category responses whereas later samples yield
increasingly finer-grained information about the cue. The resulting
iterative recognition of cues shares many characteristics with the
robust psychological phenomenon of "basic levels": within a
hierarchically nested set of categories such as "animal-bird-robin",
there is a specific level of abstraction that is more readily processed
(e.g., recognized faster) than the others; "bird" in this example
(Mervis and Rosch, Ann.Rev.Psych., 1981). The correspondence raises the
possibility that aspects of this psychological phenomenon may arise
from fundamental physiological mechanisms in primary sensory cortex.
-------------------------------------------------------------------
(10) ICONS AND TEMPORAL PATTERNS:
A DYNAMIC CONNECTIONIST SOLUTION TO SYMBOL GROUNDING
Harnad (1990) proposes that categories can be grounded by their direct
relationship to a physical icon of the input stimulation. The notion of
an icon has a clear meaning in the case of a visual display: it is a
pattern of activity in a field of neurons that is physically isomorphic
with the pattern of light in 2D. The standard proposal for a physical
icon of TIME is physical distance. Such a model is naturally
implemented by delay lines in a network (see Lang, 1990, NN).
But delay lines are not a good model for human behavior. When subjects
are trained on a complex temporal pattern, like a random sequence of
tones, they can develop a detailed perceptual representation (Spiegel &
Watson 1981, Watson and Foyle, 1985, JASA). Some skills that should be
easy are very difficult -- eg, recognizing an absolute time interval in
the face of randomly varying intervening sounds. This should be easy
since absolute time differences are represented by weights on specific
delay lines. So edges from inputs that have a random relation to the
categorical identification should learn random weights. On the other
hand, skills that should be very difficult turn out to be easy, such as
detecting serial order of familiar patterns in the face of changes in
rate of presentation. This should be difficult since a pattern that
appears at different rates will be distributed differently across the
range of delays, and should thus be learned only slowly.
On other hand, we have been developing dynamic network models that
represent learned temporal patterns of tones as stable equilibria in
the activation space of a group of fully recurrent nodes
(Port-Anderson, 1990, Anderson-Port, 1990). These systems were trained
(with real-time recurrent learning) to recognize particular tone
sequences. They are highly resistant to noise and continue to recognize
patterns even when the rate of presentation is varied by a factor of 2
faster or slower. Watson's results and our simulations suggest that
brains do not produce an icon of auditory patterns in time. It implies
that direct contact with stimulation that is distributed in time is not
possible. Although I do not disagree with Harnad on the importance of
symbol grounding for an account of perception, apparently, the
grounding of categories does not require an `icon' in the sense that
Harnad has proposed.
-----------------------------------------------------------------
--
Stevan Harnad Department of Psychology Princeton University
harnad at clarity.princeton.edusrh at flash.bellcore.comharnad at elbereth.rutgers.eduharnad at pucc.bitnet (609)-921-7771