Title: Lecture 7: Basics of Neural Nets and Past-Tense model
1COM1070 Introduction to Artificial Intelligence
week 10 Yorick Wilks Computer Science
Department University of Sheffield www.dcs.shef.ac
.uk/-yorick
2Characteristics of von Neumann architecture
- A von Neumann machine is a sequential processor
executes instructions in a program one after
another. - Each string of symbols stored at specific memory
location. - Access to stored string via numerical address of
strings location. - Single seat of control, central processing unit,
CPU. - But brain may still be a computer, just not a von
Neumann machine, or possibly a VNM with enormous
redundancy of data.
3Ability of GOFAI and Neural Nets to provide an
account of thought and cognition
- The debate is about whether it is possible to
provide a neural computing (NN) account of
cognition, or whether we should assume that a
symbol system (GOFAI, or Good Old Fashioned AI)
is required. - Physical symbol system hypothesis it is possible
to construct a universal symbol system that
thinks - Strong symbol system hypothesis only universal
symbol systems are capable of thinking. - I.e. anything which thinks (e.g. human brain)
will be a universal symbol system. - I.e. all human thinking consists of symbol
manipulation. - I.e. only computers can think therefore we are
computers. - If its not a universal symbol system, it cant
think!
4- Pylyshyn an advocate of strong symbol system
hypothesis - Our mental activity consists of manipulation of
sentence like symbolic expressions. - Human thought manipulation of sentences of
internal mental language. - Fodors Language of Thought or Mentalese
- We can show that brain does not have a von
Neumann architecture. - But this does not disprove Strong Symbol System
Hypothesis. - Strong symbol system hypothesis says nothing
about architecture.
5- Could have symbol manipulation in parallel system
- Could have different memory system i.e.
content-addressable storage - Instead of CPU could have different parallel
control. - Brain can form representations of events in
world. - SSSH representations in brain take form of
sentence-like strings of symbols. - I.e. sentence Fido licks Brian is symbolic
representation of fact that Fido licks Brian. - Language of thought, or Mentalese
- Mentalese representations not literally English,
but like sentences, in that they have basic
vocabulary units, and form in which they are
arranged determines their meaning.
6- Advantages of mentalese
- Provides an account of beliefs, intentions and
doubts etc. these are also expressed in terms of
mentalese sentences. - Provides an account of productivity and
systematicity. - Human language is productive, (no limit to number
of sentences we can produce). - And systematic (if you can say John loves Mary,
you can also say Mary loves John) - Human thought productive and systematic because
it relies on Mentalese, which is productive and
systematic. - Origin of all this in Chomskys generative
linguistic theories.
7- Neural Computing can it form an alternative to
SSSH? - Is it possible to provide a connectionist account
of thought, or is it as SSSH advocates would
claim, impossible? - Symbols Subsymbolic Hypothesis versus Strong
Symbol system hypothesis - Symbol System Hypothesis (Newell and Simon, 1976)
- ..a physical symbol system has the necessary and
sufficient means for general intelligent
action.. - A symbol designates something.
- A symbol is atomic (cannot be broken down
further). - E.g. elephant designates an elephant. Or P
could designate elephant. Or 01100 could
designate elephant, but with no interpretation of
the 1s and 0s.
8- Compositional symbol compound symbol which has
meaningful parts, whose overall meaning is
determined by meaning of those parts. - E.g. sentences of natural language.
- The kangaroo jumped over the elephant.
- Distinction between symbol types and symbol
tokens - E.g. in AMBIGUITY, 9 letter tokens, 8 letter
types. - Same symbol type can be realised in many
different ways. - Symbol is both (a) a computational token, and (b)
it designates something, i.e. it is a
representation.
9Connectionism subsymbolic hypothesis
- (see Smolensky, 1988).
- Symbols can be broken down, computation takes
place at subsymbolic level. - Connectionist representations distributed
represent-ation. - Distributed representation pattern of activity
over several nodes. - E.g. in a distributed representation of
elephant, elephant is represented by
distributed represent-ations over several nodes. - Thus, there is not an atomic symbol for elephant.
10- So connectionism rejects Strong Symbol System
Hypothesis - To complicate matters, one can have localist
neural networks, where concepts are represented
by single nodes. (e.g. node which represents
elephant). Localist connectionism assumes its
symbols given and may be compatible with the
SSSH. - This discussion applies to distributed
(non-localist) representations.
11- Distributed representations subsymbolic
computation. - E.g. representing letter A
- Localist scheme, single unit for A, one for B
etc. - Or (in a subsymbolic system) letters represented
as patterns of activity across 78 units. E.g. A
units 1,2, and 3, B units 4,5, and 6 etc. - Individual units stand for features of letters.
Thus letter A will be joint activity of various
features it contains. So letter E will share
several features with letter F. Thus similarities
and differences among items are reflected in
similarities and differences among
representations.
12- Symbolic-Subsymbolic distinction.
- In a symbolic system the computational level
coincides with the representational level. In a
subsymbolic system, the computational level lies
beneath the representational level. - Symbol is both (a) a computational token, and (b)
it designates something, i.e. it is a
representation. - But in subsymbolic Connectionism representations
across several units, but computational tokens
are units. - But does the same issue reappear? Subsymbolic
computation may not assume the symbols but what
about their features--e.g. horizontal strokes for
E and F??
13Symbolic criticisms of connectionism
- Arguments in favour of Strong Symbol System
Hypothesis - Fodor and Pylyshyn (1988)
- Argument that connectionism is inadequate as a
representational system. - Focus on issues of compositionality and
systematicity - Fodor and Pylyshyn (1988) argue
- Compositionality and structure sensitivity are
necessary to support cognitive processes. - Only classical representations exhibit
composition-ality and structure sensitivity
(systematicity) - Only classical representations can support
cognitive processes.
14- Compositionally concatenative representations
Where complex representations are composed from
primitive tokens combined concatenatively - Molecular representations can be formed out of
constituents, and can be manipulated in accord
with syntactic rules. - E.g. the kangaroo jumped over the elephant
- If you can understand this, you can also
understand a sentence which has the same words in
a different order. (Systematicity). - E.g. The elephant was jumped over by the
kangaroo - or
- The elephant jumped over the kangaroo
15- Two sentences are composed of same elements.
Constituent elements manipulated according to
syntactic rules. - But, according to Fodor and Pylyshyn,
connectionist representations cannot be
manipulated like this. - FP cannot compose simple connectionist
representations into more complex representations - Much of this argument comes down to the role of
tree structures or hierarchies, which are needed
to express syntactic relationships --- SSSH
people say trees cannot be learned by
connectionist systems.. - But Jordan Pollack showed in 1986 that they can
(up to a point anyway)--this argument is very
like a return of the XOR argument at a higher
level.
16Connectionist Counter-Arguments
- Distributed representations nonsymbolic and
continuous (developed over hidden units). - But van Gelder (1990), can have functional
compositionality - .. We have functional compositionality when
there are general, effective and reliable
processes for (a) producing an expression given
its constituents, and (b) decomposing the
expression back into those constituents - Many nets can do (a) and (b)
- For example, net can be trained to structurally
disambiguate sentences. - John saw money with the telescope.
- (John (saw money) (with the telescope)).
17- Complex representations of input sentences
developed over hidden units. - Then these representations decomposed into
required output (ii). - Thus, fully distributed representations carry
information about syntactic structure of inputs,
without being syntactically structured. - I.e. they demonstrate a functional
compositionality - by moving from structures to their components and
back again.
18- .
- But can distributed representations permit
structure sensitive operations? (systematicity) - (e.g. changing from active to passive).
- Fodor and McLaughlin (1990) to support
structure sensitive operations, representations
must contain explicit tokens of original
constituent parts of complex expression. - But can show that connectionist representations
can have property of systematicity (permitting
structure sensitive operations).
19- Chalmers (1990), trained a connectionist net to
transform representations of active sentences, to
passive sentences. - Developed representations for both active and
passive sentences. Used RAAM nets (Pollack, 1990)
to do this. - Took the fully distributed representations from
the RAAM nets, for both active and passive
sentences. - Trained a net to translate from active sentences
to passive sentences. - Training complete could input active sentences,
extract representation, and translate that into
the representation for a passive sentence, which
could be decoded into the passive sentence.
20- Also generalised to sentences that were not part
of the training set. - I.e. Connectionist representations did permit
structure sensitive operations, without being
decoded into symbols.
21Summary of representation argument
- Claimed by Fodor and Pylyshyn that
- Compositionality and structure sensitivity are
necessary to support cognitive processes. - Only classical representations exhibit
composition-ality and structure sensitivity
(systematicity) - Only classical representations can support
cognitive processes. - BUT
- Can demonstrate that connectionist nets show
functional compositionality. Although
representation do not contain tokens, as symbolic
representations do, these tokens can be obtained
from the representations.
22- E.g. Can train nets to disambiguate sentences
like the following. Intervening hidden unit
representation does not contain explicit tokens,
but output does. - John saw money with the telescope.
- (John (saw money) (with the telescope))
- Can also show that connectionist representations
can support structure sensitive operations
(systematicity). - Chalmers (1990) Translating from active to
passive sentences using connectionist distributed
representations. - So counter-argument against claims that neural
nets cannot provide an account of cognition. So
Neural Computing can provide an alternative to
the Strong Symbol System Hypothesis.
23- But has yet to be shown if can provide a
connectionist account of all aspects of thought. - Smolensky, (1988) .. It is far from clear
whether connectionist models have adequate
power to perform high level cognitive tasks.. - Connectionism Does well at accounting for
low-level aspects of cognition e.g. movement,
pattern recognition. - But Beliefs? Intentions?
- Can Neural Computing provide an account of
Direct Conscious Control? (Norman, 1986) - E.g. consciously planning what to do,
introspecting about our thoughts, holding
beliefs, making logical inferences, early stages
of learning a skill.
24- Possible that brain best modelled in terms of
hybrid system - Connectionist account of lower level processes
- Symbolic account of higher level processes.
25 Adaptive Behaviour, and Symbol Grounding
Revisited
- Three different approaches to Artificial
Intelligence, two of which we have already
encountered many times - Symbolic AI or Traditional AI or GOFAI (Good Old
Fashioned AI) - Neural Computing or Connectionism or Parallel
Distributed Processing - Adaptive Behaviour or Behaviour-based Robotics
- Adaptive Behaviour
- Rodney Brooks, at MIT Artificial Intelligence
Lab. (see reference to Brooks on web page for the
course) - Brooks, R.A. (1991) Intelligence without Reason.
MIT AI Lab Memo 1293, April 1991
26- An example Allen, a reactive robot. (named after
Allen Newell?) - Sonar sensors, and odometer to keep track of
distance travelled. - Controlled by cable from off-board special
purpose computer. - Lowest level reactive layer used sonar readings
to keep away from moving and static obstacles. -
if an obstacle is close, instead of bumping into
it, stop. - Second level random wandering. Every 10 seconds,
generate a movement in a random direction. - Third level Look for a distant place, and move
towards it. Odometry can be used to monitor
progress. - Three layers made it possible for robot to
approach goal, whilst avoiding obstacles.
27- Goal subsumption switching control between the
modules is driven by the environment, not by a
central locus of control. - Robot heads for goal until sensors pick up
information that there is an obstacle in the way.
The obstacle avoidance module cuts in. Once the
obstacle has been avoided the goal-finding module
takes control again. - Robot can move around in the environment although
it does not build, or use, any map of that
environment, and is only driven by simple
environmental cues.
28- Second example Herbert (Herbert Simon?)
- Wanders about an office environment, picking up
coke cans and returning them to start point. - Sensors infrared ports, and laser 3D data.
- Actuators motors driving wheels, and manipulator
arm with sensors. - Subsumption architecture several
behaviour-generating modules. - Modules include obstacle avoidance, wall
following, and recognition of coke cans. - Control of modules Only suppression and
inhibition between alternative modules - no other
internal communication. - Each module connected to sensors and to
arbitration network which decides which competing
action to take.
29- Description of Herbert in action
- When following a wall, Herbert spots a coke can.
The robot locates itself in front of the can.
The arm motion is then begun - when can is
detected with sensors local to the arm, it is
picked up. - Advantages naturally opportunistic. If coke can
put right in front of Herbert, can collect it and
return to start, since no expectations about
where coke cans will be found. Can find coke
cans in a variety of locations, even if never
found there before. - But.
30(No Transcript)
31Behaviour-based Robotics
- Idea of building autonomous mobile robots.
- New approach, where robots operate in the world,
and use highly reactive architectures, with no
reasoning systems, no manipulable
representations, no symbols, and totally
decentralized computation (Brooks, 1991) - I wish to build completely autonomous mobile
agents that co-exist in the world with humans,
and are seen by those humans as intelligent
beings in their own right. I will call such
agents Creatures... (Brooks, 1991) - Brooks, R. (1991) Intelligence without
Representation Artificial Intelligence, 47,
139-159.
32- See Elephants dont play chess, (1990 paper by
Brooks) - Brooks, Rodney A. (1990) Elephants dont play
chess. In Pattie Maes (Ed) Designing autonomous
Agents, Cambridge, Mass MIT Press. - Because elephants dont play chess, no reason to
assume they are not intelligent. - Emphasis on kind of behaviour exemplified by
elephants, rather than on more abstract human
behaviours (e.g. games, speech recognition,
problem solving).
33- A Creature must cope appropriately and in a
timely fashion with changes in its dynamic
environment. - A Creature should be robust with respect to its
environment minor changes in the properties of
the world should not lead to total collapse of
the Creatures behaviour rather one should only
expect a gradual change in the capabilities of
the Creature as the environment changes more and
more. - A Creature should be able to maintain multiple
goals and, depending on the circumstances it
finds itself in, change which particular goals it
is actively pursuing thus it can both adapt to
surroundings and capitalize on fortuitous
circumstances. - A Creature should do something in the world it
should have some purpose in being.
34- Set of principles (Brooks, 1991)
- The goal is to study complete integrated
intelligent autonomous agents. - The agents should be embodied as mobile robots
situated in unmodified worlds found round
laboratory. (embodiment). - Robots should operate under different
environmental conditions - e.g. in different
lighting conditions, when sensors and actuators
drift in calibration (situatedness). - Robots should operate on timescales commensurate
with timescales used by humans (situatedness).
35Key Topics of Behaviour-based Approach
- Situatedness
- Embodiment
- (animal or insect) Intelligence
- Emergence
36Situatedness
- A situated automation is a finite-state machine
whose inputs are provided by sensors connected to
the environment, and whose outputs are connected
to effectors. - The world is its own best model
- Traditional AI, working in symbolic abstracted
domain. - Problem solvers which are not participating in
the world as agents. - Dealing with model world - no real connection to
external world.
37- Alternative approach, to use a mobile robot which
uses the world as its own model, referring to
information from sensors rather than internal
world model. - Representations are developed which capture
relationships of entities to robot. - Situated agent must respond in timely fashion to
inputs but much information from the world.
38Embodiment
- The world grounds symbolic regress
- Embodiment Physical grounding of robot in real
world. - According to Brooks (1991), embodiment is
critical for 2 reasons. - Only an embodied agent is validated as one that
can deal with real world. - Only through a physical grounding can any
internal symbolic system be given meaning.
39Brooksian view of Intelligence
- Intelligence is determined by the dynamics of
interaction with the world - Some activities we think of as intelligent have
only been taking place for a small fraction of
our evolutionary lineage. - Simple behaviours to do with perception and
mobility took much longer to evolve. - Would make sense to begin by looking at simpler
animals. - - looking at dynamics of interaction of robot
with its environment.
40Emergence
- Intelligence is in the eye of the observer
- Intelligence emerges from interaction of
components of the system. - Behaviour-based approach - intelligence emerges
from interaction of simple modules. - e.g. Obstacle avoidance, goal finding, wall
following modules.
41- Main ideas
- No central model maintained of world
- No central locus of control
- No separation into perceptual system, central
system and actuation system - Behavioural competence improved by adding one
more behaviour specific network to existing
network. Crude analogy to evolutionary
development - No hierarchical development
- Layers or behaviours run in parallel
42Criticisms?
- This approach wont necessarily lead to system
capable of more complex behaviours. A new
controller is needed for each task. - The experimenter is deciding on what modules to
add, and what environment and task the robot
should be exposed to. - not the same as
evolution. - But in terms of evolution, new behaviours and new
mental structures are learnt in response to the
environment, not added by an experimenter. - Similarly, in the development of an individual,
new representational structures are developed in
response to the environment, not added by an
experimenter.
43- It would be more impressive if the robot learnt
new behaviour modules in response to the
environment. This possibility is discussed by
Brooks (1991), but has not yet been successfully
tackled. - Emphasis in this approach on reacting to the
environment. And it is the case that apparently
quite sophisticated behaviours can result from
simple reaction to the environment. But
representations are needed for more complex
tasks. - e.g. Find an empty can and bring it back to the
starting point - requires the formation of an internal
representation corresponding to a map. Need to
provide an account of the development of
representations.
44Symbol Grounding revisited
- Traditional view the language of thought (Fodor,
1975), that The mind is a symbol system and
cognition is symbol manipulation. - Advocates of symbolic model of mind (e.g. Fodor,
and Pylysyn) argue that symbol strings capture
what mental phenomena such as thoughts and
beliefs are. - Symbol system symbols (arbitrary physical
tokens) manipulated on the basis of explicit
rules. - Rule-governed symbol manipulation is based on
syntax of symbol tokens (not their meaning).
45- Symbols can be rulefully combined primitive
atomic symbol tokens can be combined to form
composite symbol-token strings. - Resulting symbol-token strings can be given a
meaning - i.e. they are semantically
interpretable. -
- BUT approach of assuming that mind is symbol
system can be criticised - in terms of symbol
grounding.
46- A criticism of symbol systems is that symbol
system capable of passing the Turing Test will
not be a mind, because the symbols have no
semantics (meaning) (remember the Chinese Room) - From Searle (1997) The Mystery of Consciousness
- Programs are entirely syntactical
- Minds have a semantics
- Syntax is not the same as, not be itself
sufficient for, semantics - Therefore programs are not minds. QED
47- It does not matter how well the system can
imitate the behaviour of someone who really does
understand, nor how complex the symbol
manipulations are you cannot milk semantics out
of syntactical processes alone (Searle, 1997).
48Symbol grounding, as discussed by Stevan Harnard
- Harnard, S (1990) The Symbol Grounding Problem.
Physical D 42, 335-346. - Copy of paper can be obtained from
- http//www.cogsci.soton.ac.uk/harnad/genpub.html
- N.B. see the relevant websites for this course at
- http//www.dcs.shef.ac.uk/yorick/ai_course/ai-cou
rse.html
49- Computation consists of manipulation of
meaningless symbols. - For them to have meaning they must be grounded in
non-symbolic base. - Like the idea of trying to learn Chinese from a
Chinese dictionary. - Standard reply of symbolist (e.g. Fodor, 1980) is
that the meaning of the symbols comes from
connecting the symbol system to the world in the
right way. - But how could this be done?
- Harnard provides one possible solutionSymbols
need to have some intrinsic semantics or real
meaning.
50- For Harnard, symbols are grounded in iconic
representations of the world. - e.g. consider the symbol horse
- iconic representation of a horse, is a
representation of the shapes that horses cast on
our retinas (i.e. sensory surface of the eye). - From these iconic representations (many from
individual views of horses), we form a
categorical representation - that captures the
features we need to identify a horse. - Thus the name horse is grounded in iconic and
categorical representations, learned from
experience.
51- Similarly, stripes is grounded in iconic and
categorical representations, learned from
experience. - These symbols can be combined
- zebra horse stripes.
- New symbol of zebra, is built up from the
grounded representations of horse and
stripes, which gives it meaning. - Harnard is proposing a hybrid system, in which
thought is assumed to be symbol manipulation, but
the symbols are grounded in iconic and
categorical representations of the world. - The problem with all this is WHICH symbols are so
grounded (peace, courage, Hamlet?)
52- Other solutions to symbol grounding problem have
been proposed. - Essential idea is that symbols need to be given
some meaning - need for grounding in meaningful
representations, to escape from circularity of
defining symbols in terms of symbols . - 2 other (partial) solutions
- Adaptive Behaviour and embodiment
- Connectionist (neural computing)
- Both are robot-prosthetic arguments, the first
without representations and the second with
implicit ones.
53Symbol Grounding and Adaptive Behaviour
- Would make sense to have symbols physically
grounded in real world. - Embodiment
- Only an embodied agent is validated as one that
can deal with real world - Only through a physical grounding can any
internal symbolic system be given meaning. - But adaptive behaviour people dont want to have
symbols, grounded or not.
54- Suggests a new approach to grounding symbolic
representations - but as yet no clear account of
how symbols might emerge due to such interactions
with the real world. - Emphasis in work on behaviour-based robotics has
been on behaviour without representation. - New approach, where robots operate in the world,
and use highly reactive architectures, with no
reasoning systems, no manipulable
representations, no symbols, and totally
decentralized computation. (Brooks, 1991)
55Symbol Grounding and Neural Nets
- Alternative idea symbols are grounded in
connectionist representations. - Connectionist symbols distributed
representations patterns of activation across
several units. - Connectionist symbols have internal structure.
They are not meaningless in the same way that
atomic symbols are. - This is a persuasive argument made by Chalmers
(1992) - Chalmers (1992) Subsymbolic computation and the
Chinese Room. In J. Dinsmore (Ed) The symbolic
and connectionist paradigms closing the gap,
Lawrence Erlbaum Hillsdale, New Jersey. pp 25-49.
56Topics to think about
- Mycin
- SHRDLU
- PARRY
- Expert systems
- Chinese Room
- Turing Test
- Weak Strong AI
57- Neural networks
- Adaptive behaviour
- Symbolic AI
- Symbol grounding