connectionism and ontological...
TRANSCRIPT
-
Peter Gaffney / PHF 2009-2010
Connectionism and Ontological Realism
(Or “This is a tank.”)
The appearance of the mirror already introduced into the world of perception an ironical effect of trompe-l’oeil, and we know what malefice was attached to the appearance of doubles. But this is also true of all the images which surround us: in general, they are analyzed according to their value as representations, as media of presence and meaning. The immense majority of present day photographic, cinematic and television images are thought to bear witness to the world with a naïve resemblance and a touching fidelity. We have spontaneous confidence in their realism. We are wrong. They only seem to resemble things, to resemble reality, events, faces. Or rather, they really do conform, but their conformity is diabolical.
Jean Baudrillard, “The Evil Demon of Images” (1987)
Representations are bodies too!
Gilles Deleuze and Félix Guattari, A Thousand Plateaus (1987)
-
PART I: Computer vision and the problem of “investiture”
The problem
When you ask somebody in the field of computer vision about the limitations of neural
networks, they are likely to tell you this story1: In the 1980s, the Pentagon implemented a plan to
equip each of its tanks with a device that would scan the visual field for enemy tanks and alert
the driver if it saw anything suspicious. The plan looked feasible enough as far as hardware goes
(a camera connected to a computer), but when it came to writing a program capable of
deciphering the complexity of the visual field in real time, they found that conventional
algorithm-based data processing was ill suited to the task. Making matters worse, the device
would have to be capable of detecting enemy tanks hiding behind trees or other objects in the
visual field. So they opted for an artificial neural network, a type of program that is designed to
mimic the way biological neurons approach the task of pattern recognition (and other perceptual-
cognitive functions). We will see in a minute how this differs from algorithmic symbol-based
programs, and how neural networks gain a key advantage when employed in computer vision. In
the case of the tank-detection program, it is important to note that neural networks, unlike
conventional programs, do not start out with a complete understanding of how to execute a
particular task; they have to be trained. So the programmers took 100 photos of tanks hiding
behind trees and another 100 of trees without tanks. They fed half of all the photos into the
neural net and put the other half aside for testing the program when training was complete. Each
time a photo was presented to the net, it was asked if there was a tank hiding behind a tree.
1 I based my version of the story on the article “Neural Network Follies” by Google software engineer Neil Fraser. The same article is cited in Priddy and Keller, Artificial Neural Networks: An Introduction (Bellingham, Washington: The International Society for Optical Engineering, 2005), and Crochat and Franklin, “Back-propagation neural network tutorial” (http://ieee.uow.edu.au/~daniel/software/libneural).
Gaffney, "Connectionism and Ontological Realism" Page 2
-
Initially, the program was only able to produce random guesses, since it did not even know what
it was supposed to do, let alone what features might make a tank visually distinct from a tree.
Distinctions like these (the task itself, in essence) must be “taught” to the neural net. This is done
by changing the way excitement flows from one “neuron” (or unit) to another⎯that is, by
modifying the “weights” of all connections within the network until it consistently produces the
desired distribution of electrical excitement for a given input. Eventually, the program was able
to identify every photo in which a tank was hiding behind a tree. It performed equally well when
it was given the 100 photos initially set aside (testing stage). But the programmers wanted to be
sure, so they returned to the field and took 100 more photos. This time, however, the answers
that the neural net generated appeared to be completely random. What the programmers had
failed to consider is that the initial 100 photos with tanks had been taken on a sunny day and all
the others when it was cloudy. Guided by no symbolic representation on which to base its
deductions during the training stage, the neural net had been looking all along at the color of the
sky. As Neil Fraser puts it, “The military was now the proud owner of a multi-million dollar
mainframe computer that could tell you if it was sunny or not.”
What does this story tell us? At first glance, it seems to illustrate a fundamental flaw with
neural nets: we just do not know what they learn.2 But what about the programmers? They had
neglected to control a key variable in the training period (weather) that just happened to coincide
with the target pattern (hidden tanks). This may seem like an obvious error⎯though, apparently,
not an uncommon one3⎯, but we can imagine other, subtler patterns that a programmer might
2 It is worth looking at Neil Fraser’s exact words here: “This story…is a perfect illustration of the biggest problem behind neural networks. Any automatically trained net with more than a few dozen neurons is virtually impossible to analyze and understand. One can’t tell if a net has memorized inputs, or is ‘cheating’ in some other way.” 3 Rolf Lakaemper, a computer vision and robotics expert who first brought this story to my attention, added that programmer errors of this magnitude happen all the time.
Gaffney, "Connectionism and Ontological Realism" Page 3
-
inadvertently feed to a neural network during the training period, patterns that would not
necessarily be as obvious as the weather. We can even imagine patterns that emerge as a
determining factor in the regular behavior of a neural net without ever becoming detectable to
the programmer⎯until some catastrophic later moment when the two patterns suddenly diverge.
It is plausible, for example, that birds partially hidden in the photos of trees were not present in
the ones with tanks. Such a coincidence might initially help a neural net produce the desired
output, until the day when birds are no longer afraid of tanks, or some combination of season and
terrain means there are no birds to begin with.
But this is precisely what is so striking about neural nets. At no point in the training
period can we say that they learn to recognize the target object⎯indeed, we cannot say they
learn to recognize any object at all, not even the color of the sky, so long as we define these
things first and foremost as symbolic representations, each one with a finite sets of attributes.
What a neural network lacks is not the capacity to react consistently to certain patterns, but the
power to treat intersecting patterns as discrete entities. We have already gone too far when we
say that the tank-detection program can tell us if it is sunny or not, since it could just as well be
some other pattern that intersects with fair weather, one that is generally present⎯but not limited
to⎯the collection of attributes that we call a “sunny sky.” The most we can say about a neural
network, in this sense, is that its numerous connections can be modified to distribute electrical
impulses in a certain pattern every time it encounters some other pattern (but which?) in the real
world. By the same token, we cannot say that responses produced by the tank-detection program
during the final phase of testing were random in any sense of the word. They only appear random
from the point of view of a programmer who is tracking the features of a different object.
Gaffney, "Connectionism and Ontological Realism" Page 4
-
From basic neural nets to adaptive-subspace self-organizing maps (ASSOM)
In their basic conception, neural networks designed for pattern-recognition do not differ
greatly from conventional feature extraction and mapping. Early designs for these networks
provided a new platform for processing complex data, one that led to a new paradigm in
cognitive science and artificial intelligence, namely connectionism. But it was not until the early
1980s that Teuvo Kohonen, a Finnish physicist working on the problem of adaptive learning in
neural nets, developed a method for taking full advantage of its novel characteristics.
Connectionism gets its name from the numerous synaptic-like connections that join one layer of
units to another in a neural net. In a simple “feed-forward” network (fig. 2), units on an input
layer are activated in a certain pattern (+1, -1, -1, +1). Each one of these units sends a charge to
each other unit on the output layer, but the strength of that charge (the amount of activation)
depends on the weight of each connection. It is by modifying these weights that a programmer
can “train” a neural net to produce a desired pattern when it is given a particular input. Since it is
the programmer who changes these weights (according to a technique called “backpropagation”),
Gaffney, "Connectionism and Ontological Realism" Page 5
-
it is not clear at first why the proper weights are not simply calculated in advance and
incorporated into the design of the machine. But this can be explained in part by the fact that
many connectionists consider their research as a basis for bridging the conceptual gap between
artificial intelligence and cognitive science. Neural nets provide a functional model that explains
some of the same adaptive properties as real neurons in the brain, and therefore bring new insight
to key issues in both fields.
But is there any practical advantage to be gained by reaching the appropriate
configuration of weights over many (sometimes thousands) of training cycles or “epochs”? One
answer has to do with the capacity for neural networks to memorize and implement many
different patterns. A simple two-layered network designed by James McClelland and David
Rumelhart was able to match four eight-digit patterns using eight input units and eight output
units (Bechtel and Abrahamsen 1991, 2002: 93). This is remarkable considering the similarities
between patterns, which consisted only in positive and negative integer activation values, as in
the example above. Even more impressive is the network’s ability to cope with distorted input
data. William Bechtel and Adele Abrahamsen built a two-layered network but simulated
distortion by adding a random value between 0.5 and -0.5 to the activation of each input. In only
50 epochs, the network produced the appropriate output for each of the four patterns with a
margin of error of less than 0.2. In other experiments, they reversed the sign of one of the units
or gave it an input pattern it had never seen before. In both cases, the neural net performed well,
with nearly error-free output in the first case and plausible generalized output in the second (i.e.
it chose the pattern from the four it had learned that most resembled the new pattern).
Examples like these demonstrate how complex pattern-recognition tasks can be carried
out with relatively few units simply by shifting the work of computation to the connections
Gaffney, "Connectionism and Ontological Realism" Page 6
-
between them. But it is difficult to imagine how this principle could be directly applied to real-
life pattern-recognition devices like the one designed to detect tanks. For one, we have been
looking at models for neural networks in which the units are all uniform, and uniformly
distributed. How would such a network approach the task of selecting out those features that
belong to a tank or any other object, so that it could distinguish it from other objects in the visual
field? The easiest solution is to design each input neuron to detect a different feature and then
relay it to another layer where it can be incorporated into higher representations: this is called
“feature extraction.” In 1959, Oliver Selfridge designed a feature-extraction and processing
principle called “pandemonium” that is a forerunner of modern Optical Character Recognition
(OCR) software. In the diagram below (fig. 3), we see how competition among the feature
“demons” leads to correct output.
Gaffney, "Connectionism and Ontological Realism" Page 7
-
As mentioned earlier, Kehonen’s approach to pattern-recognition in neural nets, called
adaptive-subspace self-organizing maps (ASSOM), differs in several ways from the examples
above:
(1) Input units are not set up in advance to serve as feature-specific (invariant)
detectors or “demons”; in their initial state, these units are variable and
uniform (like the ones in fig. 2). It is only after many epochs in the training
period that the neural net begins to acquire the capacity to detect micro-
features in a visual field.
(2) Training is unsupervised; this means that programmers have no part in changing
the weights of connections; instead, ASSOM nets are designed with several
layers of self-organizing units whose only job is to record patterns they receive
from the input layer. Weights are changed as a result of competition among
these subspace representations.
(3) This “winner takes all” approach to competitive learning also requires that units
have the capacity to exert an inhibitory effect on other groups of subspace
units. This habituation to regular patterns in its environment allows the neural
network to build successively more complex maps, but it also means that some
patterns will be filtered out by more competitive representations. This is meant
to simulate the way perceptual-cognitive faculties are able to identify the same
patterns when they are partially hidden or transformed with respect to
contingencies in the visual field. (Kehonen 1995, 2000, Kehonen 2003)
Kohonen’s explicit purpose in implementing the principle of self-organization is to avoid any
model in which the mechanism for transforming patterns into representations might be based on
Gaffney, "Connectionism and Ontological Realism" Page 8
-
a priori representations: “What we aim at is a model of experimental data, not of the process
variables” (2003: 1178). This may sound like a subtle distinction, but we have seen that it can
mean the difference between a device that detects enemy tanks and one that just connects
ineffably with the world.
The symbolic approach
It will be useful to consider how a conventional symbol-based processor would handle the
same pattern-recognition task as the tank-detection net before moving on to some conclusions
about what this means for a new “ontological” definition of perception. There are two questions
I’d like to pose in this part of my paper: (1) How do symbol-based processors work?; (2) What
can conventional principles of pattern recognition tell us about the (still ambiguous) notion of
symbolic representations, and what does this have to do with what we will have reason to call
“investiture and decoding” (or Baudrillard’s “diabolical conformity”)? In its most basic
conception, a conventional computer program functions by manipulating symbols according to
an algorithm or sequence of instructions. In the case of a program designed to identify an enemy
tank, such an algorithm might proceed from one proposition to another following the decision
tree shown below (fig. 4). From the initial condition (transmission of new data from camera to
computer), data flows through a series of “if…then…” propositions that model the object of a
tank and lead the program to the appropriate safe/alert response. The diagram in figure 4 is a
rather simplified version of what might be required to complete the task of tank detection, but we
could always add more propositions to the sequence.
Gaffney, "Connectionism and Ontological Realism" Page 9
-
Rather than leading each part of the data through a series of commands, it might be better to
cluster it into different tasks. Some of these might compare new features to previous maps in
order to make subtler deductions. The diagram below (fig. 5) shows how an actual program
enables a mobile robot to build maps of its environment autonomously. This feature is called
“simultaneous localization and mapping” or SLAM:
Gaffney, "Connectionism and Ontological Realism" Page 10
-
What is important in the first example is that any instruction added to the algorithm to further
qualify a “Yes” statement must be inserted between two instructions in the sequence. Similarly,
an instruction or series of instructions used to qualify a “No” would either proceed to a “Safe”
output or back to the next instruction in the main sequence. Propositional logic thus shares its
essential structure with the syllogism, such as the famous “All men are mortal; Socrates is a man;
therefore Socrates is mortal.” So long as the constituent propositions are true, the final deduction
must also be true; it is a “truth-preserving” device. Of course, we don’t need a data processor to
deduce that Socrates is mortal; this is already built into our higher representations of Socrates, if
not immediately given in the attribute which tells us that Socrates is already dead. But not all
syllogisms are so straightforward. We can add as many premises to the syllogism as we want,
Gaffney, "Connectionism and Ontological Realism" Page 11
-
creating more stipulations in a chain of reasoning (one that goes beyond the simple task of
pattern-recognition). Here’s a more complicated example:4
Instruction in natural language variables substituted If industries are to be kept going, if B there must be a steady supply of oil: then C If the United States is to prosper, if A its industries must be kept going: then B therefore if the United States is to prosper (therefore) if A there must be a steady supply of oil. then C In order to secure a steady supply of oil, if C the United States must go to war: then D therefore if the United States is to prosper, therefore if A it must go to war. then D therefore if it does not go to war, therefore if not D the United States will not prosper. then not A
SLAM is another algorithm-based program, but the flowchart above (fig. 5) does not
describe the attribution of symbolic values to objects in the visual field. It only shows how the
robot extrapolates its position by comparing new visual data to an existing map (localization),
and to modify the map whenever it diverges from the extrapolated position (mapping). But let’s
say we were hired to modify this robot for an urban combat environment where it must identify
and destroy enemy combatants. We would be in the same position as the tank-detection
programmers, but with a much greater imperative to assign and manipulate the proper symbolic
values for each target object. So long as we were using algorithm-based processing, we would
need to develop a series of compounding stipulations similar to the one above (if A and B and C
and D and E… then X, but only if Y = Z, etc.). This time, there would be a much greater burden
4 This example is a modified version of the one found in Chapter 83 of R. W. Jepson’s “Clear Thinking” (http://www.ourcivilisation.com/smartboard/shop/jepsonrw/chap83.htm).
Gaffney, "Connectionism and Ontological Realism" Page 12
-
on the side of the programmers, if not on the program itself, to determine the thresholds between
targeted and non-targeted objects. Is an object in the visual field acting hostile? Is it hiding a
weapon? Is it harboring insurgents? What is at issue here is not the effectiveness of feature
extraction vis-à-vis some partially obscured object in the visual field, but the capacity for
programmers to invest an image with whatever representations will be necessary to complete a
given task. The real question is about human interest and context, not computer vision.
In War in the Age of Intelligent Machines (1991), Manuel DeLanda explores some of the
troubling implications of this issue in the case of military automata:
The PROWLER [fig. 6]…is a small terrestrial armed vehicle, equipped with a primitive
form of “machine vision” (the capability to analyze the contents of a video frame) that
allows it to maneuver around a battlefield and distinguish friends from enemies. Or at
least this is the aim of the robot’s designers. In reality, the PROWLER still has difficulty
negotiating sharp turns or maneuvering over rough terrain, and it also has been deployed
only for very simple tasks, such as patrolling a military installation along a predefined
path. We do not know whether the PROWLER has ever opened fire on an intruder
without human supervision, but it is doubtful that as currently designed this robot has
been authorized to kill humans on its own…For now, the robot simply makes the job of
its human remote-controller easier by preprocessing some of the information itself, or
even by making and then relaying a preliminary assessment of events within its visual
field. (1991: 1)
In its practical application, the computer vision of PROWLER and other automata of its era is
little more than a surveillance camera. But DeLanda goes on to point out that in the military
Gaffney, "Connectionism and Ontological Realism" Page 13
-
application of artificial intelligence, “it is precisely the distinction between advisory and
executive capabilities that is being blurred” (2). He gives the example of a war game designed to
help the Pentagon learn which sequence of actions comprise the winning strategy; the problem is
that “SAM and IVAN [the simulated players of the game]…do not have any problem triggering
World War III” (2).
Of course, this problem is not limited to algorithm-based programs, nor does it first
emerge at the level of programming. As we have already seen, the design of any pattern-
recognition device has its starting point in a network of representations that belong to the
designers, to people further up the chain of command (those who conceive and finance the
project), and eventually to a heterogeneous field of social codes which determine the flows of
knowledge, investment capital, raw materials, and so on. Likewise, computer-generated output is
only meaningful in relation to a broader set of representations and to those who are interested in
its results. There is no such thing as a computer that sees, a computer that knows, but this has
Gaffney, "Connectionism and Ontological Realism" Page 14
-
more to do with the agency behind it than anything in its design. Such is the predicament John
Searle indicates when he writes: “What [a computer] does is manipulate formal symbols. The
fact that the programmer and the interpreter of the computer output use the symbols to stand for
objects in the world is totally beyond the scope of the computer” (1980: 437). Searle takes the
more extreme view that a computer program is not capable of processing actual images, only
other, non-computer symbol systems: the designations coded in a natural language, for example,
together with the underlying cultural matrix that sanctions how these designations are made.
What we have been calling pattern recognition thus consists in projecting meaningful content
onto a visual field and then interacting with a social and physical environment on that basis.
Programmers may very well have reason to be optimistic about the growing power and
efficiency of computer vision technology. But this coincides with growing concern among Searle
and others that there are ideological consequences of symbol-manipulating technology. By
confusing intelligence and “investiture,” the symbolist approach to cognitive science serves only
to reify our representations by pretending to find them “in the outside world.”
Seeking and finding: the notion of “investiture”
It is precisely in this sense that Victor Burgin5 speaks of a “photographic paradox”
(Thinking Photography). It is worth quoting these texts at length here, since they illustrate how
closely the practical problems we have been discussing resemble⎯because they derive
from⎯the more elusive mechanism of investiture:
5 It should be mentioned that many of Jean Baudrillard’s works (The Evil Demon of Images, The Gulf War Did Not Take Place, among others) trace the same development of a “diabolical conformity” in the mediatization of the event. I have chosen to follow Victor Burgin’s statement of the problem because it so closely resembles the pattern-recognition problem.
Gaffney, "Connectionism and Ontological Realism" Page 15
-
The structure of representation⎯point-of-view and frame⎯is intimately implicated in
the reproduction of ideology (the “frame of mind” of our “points-of-view”). More than
any other textual system, the photograph presents itself as “an offer you can’t refuse.”
The characteristics of the photographic apparatus position the subject in such a way that
the object photographed serves to conceal the textuality of the photograph
itself⎯substituting passive receptivity for active (critical) reading. When confronted with
puzzle photographs of the “What is it?” variety (usually, familiar objects shot from
unfamiliar angles) we are made aware of having to select from sets of possible
alternatives, of having to supply information the image itself does not contain. Once we
have discovered what the depicted object is, however, the photograph is instantly
transformed for us⎯no longer a confusing conglomerate of light and dark tones, of
uncertain edges and ambivalent volumes, it now shows a “thing” which we invest with a
full identity, a being. With most photographs we see, this decoding and investiture takes
place simultaneously, unselfconsciously, “naturally”; but it does take place⎯the
wholeness, coherence, identity, which we attribute to the depicted scene is a projection, a
refusal of an impoverished reality in favor of an imaginary plenitude. (Burgin 146-147)
The problem with mechanical reproductions is not that they fool the eye or cause us to confuse
reality with its likeness, but that they generate the comforting illusion, as a function of their
realism, that symbolic representations spring readymade from the visual field. This is what
makes photographs so persuasive. Similarly, if a computer is able to decode an image, if it is able
to provide an appropriate answer every time it is asked “What is it?” this is only because the
visual field has been prepared in advance to respond like a code. The presence of an object is not
Gaffney, "Connectionism and Ontological Realism" Page 16
-
“recognized,” any more than a pattern is “decoded.” Rather, these functions⎯truth itself, in
essence⎯arrive later on the scene, as the result of an investment at the level of the symbolic.
Symbolist theories of human intelligence, taking their cues from technological advances in the
field of computer vision, thus proceed by treating the world as if it conformed to the rituals of
modern visual culture. What first confronts the eye as a confusing jumble of dark and light is
subsequently brought into good order by the logical play of representations. But this is not the
same thing as vision. It is only the happy coincidence of meaning and presence, a game of hide
and seek like the one Nietzsche describes in his challenge to the concept of truth: “When
someone hides something behind a bush and looks for it again in the same place and finds it
there as well, there is not much to praise in such seeking and finding. Yet this is how matters
stand regarding seeking and finding ‘truth’ within the realm of reason” (2000: 157).
We should not view such criticisms as mere philosophical wrangling. After all, it is by
the sanction of Allen Newell’s “unified theory of cognition”6 that SOAR, the symbolic cognitive
architecture that drives most of DARPA’s current designs for automata, has begun to incorporate
what DeLanda calls “executive capabilities.” Recent thinking on unmanned weapons points to
the tactical and psychological advantage to be gained by deploying a multitude⎯or
“swarm”⎯of smaller automata against an enemy. DARPA’s newest generation of unmanned
weapons will thus combine automatic targeting recognition with other algorithms meant to
mimic the way birds move in flocks and ants forage for food (Singer 2009: 232-233). As one
DARPA researcher puts it, these swarms might eventually reach the size of “zillions and zillions
of robots” (234). What is troubling about DARPA’s multitude of tiny robots is that each one will
6 It is notable that Newell argues strongly in favor of the view that symbol systems have the necessary and sufficient means for general intelligent action (“physical symbol system hypothesis”), but that “any theory of what Soar will learn must occur within a larger theory of the environments Soar inhabits, how tasks arise, and how they get defined—in short, a total theory of Soar as an agent in the world” (1994: 191).
Gaffney, "Connectionism and Ontological Realism" Page 17
-
really have to see. Unlike the PROWLER and other robots of its generation whose deductions
were subject to the discretion of a human operator, the sheer numbers of “Proliferated
Autonomous Weapons” (or PRAWNs) will make this kind of intervention impractical. “Rather
than a controller back in a mothership furiously trying to point and click at which target to
hit…the autonomous swarm would just figure it all out on its own” (234).
There is no need to reconstruct here any of the possible doomsday scenarios that such
automata bring to mind. What I would like to propose instead, in light of this notion of
investiture, is a hypothesis about the ideological tension between two moments that characterize
the social dimension of human perception, and that now plague its technological counterparts
(including many neurally-inspired processors). In the first moment, we find the profoundly
creative act that enables the mind, or perhaps a “collective mind,” to invest the visual field with
an order of its own. Burgin sees this as the projection of a pre-existing symbolic order onto a
chaotic collection of sensory data. We would do better, however, to consider it in broader terms
as a connection between two or more patterns, with the result that perceiver and perceived are
joined together in a determinate and productive relationship. We may even say, following the
example of such thinkers as Henri Bergson and Gilles Deleuze, that this simple intersection of
patterns constitutes the “actualization” of a world together with an intelligence in it7⎯so that the
two sides, perceiver and perceived, are precisely what this relationship produces. (This is a rather
confusing formula, and I will spend the second half of the paper trying to sort out its
implications). If we look closely at the examples above, we see a second moment as well, this 7 I am referring here to Bergson’s notion that “The more consciousness is intellectualized, the more matter is spatialized [Plus la conscience s’intellectualise, plus la matière se spatialise]” (1959). This notion is at the center of the Bergsonian method of intuition. Deleuze and Guattari follow the same formula when they write that “it is by slowing down that matter, as well as the scientific though able to penetrate it with proposition, is actualized” (1994). In his work on the connection between Bergson and Deleuze (Signature of the World), Eric Alliez uses a more explicit formulation: “The diversion [détournement] of Bergosnian (immediate) intuition is prepared both before and beyond…the realism-idealism opposition, when the act of knowing tends to coincide with the act that generates the real” (2004).
Gaffney, "Connectionism and Ontological Realism" Page 18
-
time characterized by a certain ideologically motivated sleight of hand. Here, the process of
actualization that creates both intelligence and world is carefully hidden from view. What was
actively produced in the first moment now appears to come readymade from an “outside” world.
Symbolism, as a philosophical doctrine, thus coincides with a visual culture (photographs,
television, cinema) that obscures the first moment by reinforcing the second. Like scientific
realism, it has the potential to actively reduce a heterogeneous complex of social codes to a
monolithic truth⎯a “royal science.”8 As Nietzsche puts it, there is not much to praise in such
seeking and finding.
PART II: Connectionism and ontological realism
Does connectionism preclude representation?
The objection might be raised that the treachery of images is a legitimate concern within
the limits of culture studies, perhaps even the other social sciences, but has nothing to do with
computer vision or science proper. The very fact that an automaton can interact with its
environment, the simple fact of applied science, should be taken as sufficient proof that task-
oriented symbolic knowledge really works. This is effectively the argument made by Newell and
Simon in their defense of the physical symbol hypothesis: “An expression [composed of
symbols] designates an object if, given the expression, the system can either affect the object
8 This concept is mapped out by Deleuze and Guattari in A Thousand Plateaus: “The ideal of reproduction, deduction, or induction is part of royal science…and treats differences of time and place as so many variables, the constant form of which is extracted precisely by the law…[royal science] implies the permanence of a fixed point of view that is external to what is produced” (372).
Gaffney, "Connectionism and Ontological Realism" Page 19
-
itself or behave in ways dependent on the object”1 (1976: 116). What right do we have to
question the truth-value of symbolic representations? On what grounds can we, like Burgin vis-à-
vis the photograph, describe them as no more than the effects of an imaginary plenitude? These
are reasonable objections, and we will have no reason to disagree with them.
The problem with symbolic knowledge is elsewhere. We spoke too soon when we said
that it is too full, that it recognizes too many “things” in the visual field. The problem is just the
opposite. In the symbolist conception, intelligence is reduced to a mechanism for verifying the
identity of being (X=X). It is, from this point of view, an epistemological machine, guided by its
truth-preserving structure and characterized by a passive receptivity rather than an active
engagement of patterns in the visual field. This is quite different from the case of the tank-
detection program, where the neural net failed precisely because it was active—too active. It did
not know that the color of the sky was an invalid attribute of the target object. It did not know
and it did not care: as an ASSOM-driven neural net, it was not trained to identify symbolic
representations, only to produce an alert signal in the presence of any pattern common to the
target samples used in the training period. The most we can say about such a neural net is that it
makes connections with the visual field. But it does not do this by investing visual data with pre-
existing symbolic value (even when operators use them as if they do). It would seem then, that
there are two ways of thinking about perception. On the one hand, it can involve the
reconstruction of “meaningless” visual data in a matrix of symbolic knowledge. In the field of
culture studies, Burgin considers how visual media naturalize this process, so that a photograph 1 Newell and Simon further specify that a physical symbol system consists of “a set of entities, called symbols, which are physical patterns that can occur as components of another type of entity called an expression (or symbol structure). Thus, a symbol structure is composed of a number of instances (or tokens) of symbols related in some physical way (such as one token being next to another). At any instant of time the system will contain a collection of these symbol structures. Besides these structures, the system also contains a collection of processes that operate on the expressions to produce other expressions: processes of creation, modification, reproduction and destruction. A physical symbol system is a machine that produces through time an evolving collection of symbol structures. Such a system exists in a world of objects wider than just these symbolic expressions themselves” (116).
Gaffney, "Connectionism and Ontological Realism" Page 20
-
or film conveys a reality that is already implicated in a matrix of symbols by the time it is
perceived. Thomas Kuhn proposes a similar theory on scientific breakthroughs, which, he
argues, do not augment an existing body of knowledge, but radically dissolve and reorganize it
around the discovery of new patterns formerly dismissed as statistical variance (concept of
“paradigm-induced gestalt switch”).2 Scientific observation, like all acts of perception, is
constructive, instrumental and historically specific—but is not for that matter less objective.
In each case, we are most likely to ask how symbolic representations give meaning and
presence to sense experience. But what happens when we trace our line of inquiry in the opposite
direction? What role does sense experience play in the production of representations? These are
not two sides of the same coin, but opposing interpretations of human intelligence that are
divided along the lines already discussed in this paper. On one side, symbolists like Newell and
Simon advance the claim that “A physical symbol system has the necessary and sufficient means
for general intelligent action” (1976: 116). Symbolism, for the same reasons as computationalism
in cognitive science and psychology, regards the mind as a kind of Turing machine which
interacts with the material world by performing purely formal operations on symbols. On the
other side, connectionists like Donald Norman and David Rumelhart argue that symbolic
representations and the rules which govern them do not play any necessary role in perception and
cognition, or in any other cognitive task for that matter. In the case of language acquisition, for
example, Norman and Rumelhart argue that “a system can appear to obey and follow general
rules of language even though it does not have those rules within it” (1981: 239). Their research
is particularly interesting since it provides evidence that even when we consciously use symbol
2 In The Structure of Scientific Revolutions, Kuhn writes that theoretical conflicts arising between existent and emergent scientific paradigms are “terminated by a relatively sudden and unstructured event like the [visual] gestalt switch” (122). It should be noted that Kuhn is in some ways critical of the comparison between scientific revolution and visual gestalt theory. He does not go so far, for example, as Norwood Hanson’s Patterns of Discovery: An Inquiry into the Conceptual Foundations of Science.
Gaffney, "Connectionism and Ontological Realism" Page 21
-
systems, these systems are not necessarily governed by rules (i.e. propositional logic). Neural
nets that are conditioned to perform basic linguistic tasks also exhibit some very human-like
error patterns during the learning process, such as the “U-shaped” acquisition of past tense
irregular verb forms11 (McClelland and Rumelhard 1988: 117).
DeLanda links the two sides of this debate to the famous controversy over the 29 Eskimo
words for snow.12 Do Eskimos perceive 29 kinds of snow because their linguistic categories “cut
up” reality in this way, or do the words derive from the sensual experience of 29 kinds of snow?
In other words, does diversity in a system of symbolic representations provide for diversity in
sense perception or is it the other way around? Hard line connectionists like McClelland and
Rumelhard claim that representations do not provide the conditions of possible experience. From
their point of view, there is no essential difference between linguistic aptitude and any other
pattern-association/completion faculty. Representations emerge as a secondary effect of
connections between one pattern and another; they are a symptom rather than a cause of complex
behavior—perhaps only a convenient metaphor. Others in the field question the usefulness, even 11 Children who correctly form past-tense irregular verbs in an initial stage will often cease to use them correctly while learning how to construct regular verb endings. In a third stage, the child learns that irregular verbs are a special case, and begins using them correctly once more. For example, a child who initially says “She went to school” might begin saying “She goed to school” during the initial stage of learning the regular verb form. Does this three-stage, or U-shaped, learning process occur because children learn to apply (and then over-generalize) a new rule? Or is learning governed only by exposure and usage, so that no explicit rule is ever internalized or applied? Proceeding from the second hypothesis, McClelland and Rumelhart constructed a neural network to mimic acquisition of the past tense. Results suggest that it followed the same U-shaped learning pattern as children who are learning a natural language (McClelland and Rumelhart 1988: 117). 12 “Everyone knows that Eskimos have 29 words for snow…The question is: Do Eskimos see 29 kinds of snow because they have 29 words for snow? That was the position taken by linguists in the 20th century⎯words cut out experience and in fact give form to experience; linguistic categories shape reality, at least phenomenological reality⎯and therefore their having 29 words for snow is the reason why they can see 29 kinds of snow. That’s one position. The other position would be the opposite: They have 29 words for snow because they can touch, feel, smell, build igloos with, and do all kinds of non-linguistic things with real snow, which comes in 29 kinds because there are at least 29 mixtures of the solid and the liquid….The two options remain quite different. In one, the 29 words are shaping perceptual reality literally. In the other, perception is not linguistic. Perception is multi-modal: it’s about sounds, smells, textures, images, all of which have their own modality and can be combined in different ways. It’s also about intervening causally in the world: to build igloos with that snow, to hunt over that snow, to find your way out of a snow storm…” (DeLanda 2007).
Gaffney, "Connectionism and Ontological Realism" Page 22
-
the validity, of continuing to use such a metaphor. Van Gelder, a leading proponent of dynamical
systems theory, compares cognitive faculties to Watt’s steam engine and centrifugal governor
(fig. 7): “when we see the relationship between engine [flywheel] speed and arm angle, we see
that the notion of representation is just the wrong sort of conceptual tool to apply…arm angle
and engine speed are at all times both determined by, and determining, each other’s behavior.”
This relationship is “much more subtle and complex…than the standard concept of
representation can handle” (van Gelder 1995: 353). Bechtel disagrees. New findings by
connectionists challenge our conventional notion of how representations work, but do not
preclude this notion altogether. In the case of Watt’s governor, Bechtel argues, a special
representation of speed is necessary in order to connect one system (referent) with another
(consumer): “Without the spindle arms and their linkage mechanism, the valve has no access to
information about the flywheel speed. They were inserted in order to encode that information in
a format that could be used by the valve-opening mechanism.” Representations thus allow the
state of affairs in one medium to be transmitted in a format that can be “understood” by another.
Gaffney, "Connectionism and Ontological Realism" Page 23
-
Deleuze and ontological realism
Put this way, however, there is no intrinsic difference between a representation and an
optical device or an eye, a taste bud or an antenna. One of the problems that arise whenever we
make a special case for representations is that we tend to get hung up on the question of truth. Is
our representation of the world an accurate one? How do we know? Can we know the world at
all? Such questions are of a purely epistemological value; they help us evaluate the status of
knowledge with respect to the unstable boundaries of what is. But they have limited value for
understanding the role of perception in constituting the real, that is, in explaining how vision
(and other senses) actively construct, dissolve and articulate patterns in a continuous coming to
be. If we put ourselves in the position of a self-organizing neural net as it engages patterns in its
environment, we come to the realization that there is no basis, much less practical reason, to
search elsewhere for “a world.” This activity is itself a world. Yet it comprises only one of many
worlds that periodically intersect, disperse, emerge and disappear again. We do not even have an
environment—an “outside”—only a milieu, assemblage or block of becoming to which our own
becoming is inextricably connected. The matrix of social codes identified by Burgin (and so
often vilified in culture studies) is just one more articulation in a network made up of many
worlds—not possible worlds or “monads,” but real intersections of abstract and material patterns.
As I have suggested elsewhere13, it is on these grounds that Deleuze interprets the predictive
success of scientific observation and the models that derive from it. From Deleuze’s point of
view, there is no need to question the truth or validity of such models, especially as this would
lead us once more to a merely epistemological evaluation of the problem. What Deleuze wants to
13 “Superposing Images: Deleuze and the Virtual after Bergson’s Critique of Science,” in Peter Gaffney ed. The Force of the Virtual (Minneapolis: University of Minnesota Press, 2010).
Gaffney, "Connectionism and Ontological Realism" Page 24
-
establish, starting with his monograph on Hume (Empiricism and Subjectivity) and borrowing
heavily from Bergson (Bergsonism), is an ontology of becoming that finds its root in all forms of
perception, conceived broadly as the creative force of the virtual in the process of actualizing
itself. A neural net does not see things in the outside world; rather, it introduces a very special
kind of eye, one that has the power to generate a set of representations that will articulate one
system with another—to generate a world made out of patterns that are themselves the
articulating feature of some other connection. Timothy Murphy has proposed that quantum
physical models provide a “realist ontological” framework for treating probably phenomena as
real events (1998: 213). In this paper, I have tried to show how connectionism provides the same
kind of framework for an ontological realist account of perception.
Conclusion
If a self-organizing neural net goes through the same stages as human intelligence when
confronted with the task of pattern-recognition and completion, then we must treat symbolic
representations as a strict redundancy. Surely representations exist, and there is no doubt that
they play a central role in the way we shape our world. But they are strange representations with
no beginning and no end, objective but not whole, real but not corporeal. They do not refer to
things; they are things. We will not be prevented, for that matter, from using representations in
the conventional manner (in this paper, for example). But it will be difficult to proceed on the
understanding that they approximate a world that precedes them, a world whose presence and
meaning confirms their validity, yet which remains just out of reach. This kind of scientific or
Cartesian realism14 no longer applies when the act of seeking is itself the constitutive feature of
14 With the term “Cartesian realism,” I am following the definition proposed by Joseph Margolis: “any realism, no matter how defended or qualified, that holds that the world has a determinate structure apart from all constraints of
Gaffney, "Connectionism and Ontological Realism" Page 25
-
whatever might be found. We cannot therefore speak of what a representation means, any more
than we can ask what a neural network learns. In each case, representation functions only as a
connection between two or more systems, one system organizing the other: a sunny sky system
organizing a neural net system, an incorporeal system organizing a material system, or vice
versa. A vast system of connections whose boundaries are continuously redrawn by the forces of
interest, desire and creativity: this is what we learn from the example of the tank-detection
device. There is nobody to blame for its failure; there is no failure. The human trainers could not
limit the task any more than the device could limit the means for accomplishing it. In any case,
the story is apocryphal.15
human inquiry and that our cognizing faculties are nevertheless able to discern those independent structures reliably. ‘Cartesianism’ serves as a term of art here, not confined to Descartes’ doctrine. It ranges over pre-Kantian philosophy, Kant’s own philosophy (quixotically), and over the views of such contemporary theorists as Putnam and Davidson” (2006: 194). 15 Though it is cited by several other authors (see note 1), I have found no published source which identifies the source of the story. Notably, it is not mentioned in William Bechtel and Adele Abrahamsen’s Connectionism and the Mind. Nick Bostrom and Milan M. Ćirković (Global Catastrophic Risk) write, “This story, although famous and oft-cited as fact, may be apocryphal…however, failures of the type described are a major real-world consideration when building and testing neural networks” (321).
Gaffney, "Connectionism and Ontological Realism" Page 26
-
WORKS CITED _______________________
Alliez, Eric. The Signature of the World: Or, What is Deleuze and Guattari's Philosophy? (New York: Continuum, 2004), 36.
Baudrillard, Jean. “The Evil Demon of Images,” in Clive Cazeaux ed. The Continental Aesthetics Reader (London: Taylor and Francis, 2000). Reprinted from The Evil Demon of Images. Trans. Paul Patton and Paul Foss (Sydney: Power Institute of Fine Arts, 1987), 13-34.
Bechtel, William and Adele Abrahamsen. Connectionism and the Mind, Second Edition (Malden, Massachusetts: Blackwell Publishers, Inc., 2002).
Bergson, Henri. L’Évolution créatrice (Paris: Les Presses universitaires de France, 1959), 132.
Bostrom, Nick and Milan M. Ćirković. Global Catastrophic Risk (Oxford: Oxford University Press, 2008), 321.
Burgin, Victor. Thinking Photography (London: Macmillan Education, 1982).
DeLanda, Manuel. War in the Age of Intelligent Machines (New York: Zone Books, 1991), 1-2.
--------“The Origins of Artificial Intelligence.” 2007 Penn Humanities Forum, University of Pennsylvania, Philadelphia. 7 November 2007.
Deleuze, Gilles and Félix Guattari, A Thousand Plateaus. Trans. Brian Massumi (Minneapolis: University of Minnesota Press, 1987).
--------What Is Philosophy? Trans. Hugh Tomlinson and Graham Burchell (New York: Columbia University Press, 1994), 118.
Dreyfus, Hubert L. “Artificial Intelligence,” American Academy of Political and Social Science, Vol. 412 (March 1974), 21-33.
Fraser, Neil. “Neural Network Follies,” http://neil.fraser.name/writing/tank)
Kehonen, Teuvo. Self-Organizing Maps (Berlin: Springer-Verlag, 1995, 2000).
Kehonen, Teuvo. “Self-Organized Maps of Sensory Events,” in Philosophical Transactions: Mathematical, Physical and Engineering Sciences, Vol. 361, No. 1807 (Jun. 15,
2003), 1177-1186.
Kuhn, Thomas. The Structure of Scientific Revolutions (Chicago: University Of Chicago Press, 1996).
Gaffney, "Connectionism and Ontological Realism" Page 27
-
McClelland, J. L. and David Rumelhard. Explorations in parallel distributed processing: a handbook of models (Cambridge, Massachusetts: Harvard University Press, 1988), 117.
Murphy, Timothy. “Quantum Ontology: A Virtual Mechanics of Becoming,” in Eleanor Kaufmann and Kevin Jon Heller eds. Deleuze and Guattari: New Mappings in Politics, Philosophy and Culture (Minneapolis: University of Minnesota Press, 1998), 211-229.
Newell, Allen and Herbert A. Simon. “Computer Science as Empirical Inquiry: Symbols and Search,” Communications of the ACM. vol. 19, No. 3 (March 1976), 113-126. Cited by Nils J. Nilsson, “The Physical Symbol System Hypothesis: Status and Prospects,” in Rolf Lungarella, Rolf Pfeifer, Josh Bongard eds., 50 Years of Artificial Intelligence, LNAI (Berlin: Springer Berlin / Verlag, 2007), 10.
Newell, Allen. Unified Theories of Cognition (Cambridge, Massachusetts: Harvard University Press, 1994).
Nietzsche, Friedrich. “On Truth and Lie in an Extra-Moral Sense,” in Clive Cazeaux ed. The Continental Aesthetics Reader (London: Taylor and Francis, 2000), 57.
Norman, Donald and David Rumelhart. “The LNR approach to human information processing.” in Cognition, Volume 10, Issues 1-3 (July-August 1981), 239.
Rockmore, Tom. “Incommensurability Modestly Recovered,” in Tom Rockmore and Joseph Margolis eds., History, Historicity and Science (Hampshire: Ashgate, 2006).
Searle, John R. “Minds, Brains, and Programs,” Behavioral and Brain Sciences, 3(3: 1980), 417-457.
Singer, P. W. Wired for War: The Robotics Revolution and Conflict in the Twenty-First Century (New York: The Penguin Press HC, 2009).
van Gelder, T. “What might cognition be, if not computation?” The Journal of Philosophy, 92 (1995), 345-381. Cited by Bechtel and Abrahamsen (1991, 2000: 275).
Gaffney, "Connectionism and Ontological Realism" Page 28
The problemInstruction in natural languagevariables substituted