Next: Remarks
Up: Humans and Robots
Previous: A conjecture about human
Human emotional and motivational structure is likely to be much
farther from what we want to design than is human consciousness from
robot consciousness.
Some authors, [Sloman and Croucher, 1981], have argued that sufficiently
intelligent robots would automatically have emotions somewhat like
those of humans. However, I think that it would be possible to make
robots with human-like emotions, but it would require a special effort
distinct from that required to make intelligent robots. In order to
make this argument, it is necessary to assume something, as little as
possible, about human emotions. Here are some points.
- Human reasoning operates primarily on the collection of ideas of
which the person is immediately conscious.
- Other ideas are in the
background and come into consciousness by various processes.
- Because reasoning is so often nonmonotonic, conclusions can be
reached on the basis of the ideas in consciousness that would not be
reached if certain additional ideas were also in consciousness.
- Human emotions influence human thought by influencing what
ideas come into consciousness. For example, anger brings into
consciousness ideas about the target of anger and also about ways
of attacking this target.
- According to these notions, paranoia, schizophrenia, depression
and other mental illnesses would involve malfunctions of the
chemical mechanisms that gate ideas into consciousness. A paranoid
who believes the CIA is following him and influencing him with radio
waves can lose these ideas when he takes his medicine and regain
them when he stops. Certainly his blood chemistry cannot encode
complicated paranoid theories, but they can bring ideas about
threats from wherever or however they are stored.
- Hormones analogous to neurostransmitters open synaptic gates to
admit whole classes of beliefs into consciousness. They are analogs
of similar substances and gates in animals.
- A design that uses environmental or internal stimuli to bring
whole classes of ideas into consciousness is entirely appropriate
for a lower animals. We inherit this mechanism from our animal
ancestors.
- Building the analog of a chemically influenced gating mechanism
would require a special effort.
These facts suggest the following design considerations.
- We don't want robots to bring ideas into consciousness in an
uncontrolled way. Robots that are to react against people (say)
considered harmful, should include such reactions in their goal
structures and prioritize them together with other goals. Indeed we
humans advise ourselves to react rationally to danger, insult and
injury. ``Panic'' is our name for reacting directly to perceptions
of danger rather than rationally.
- Putting such a mechanism, e.g. panic, in a robot is certainly
feasible. It could be done by maintaining some numerical variables,
e.g. level of fear, in the system and making the mechanism that
brings sentences into consciousness (short term memory) depend on
these variables. However, such human-like emotional structures are
not an automatic byproduct of human-level intelligence.
- Another aspect of the human mind that we shouldn't build into
robots is that subgoals, e.g. ideas of good and bad learned to
please parents, can become independent of the larger goal that
motivated them. Robots should not let subgoals come to dominate the
larger goals that gave rise to them.
- It is also practically important to avoid making robots that are
reasonable targets for either human sympathy or dislike. If robots
are visibly sad, bored or angry, humans, starting with children,
will react to them as persons. Then they would very likely come to
occupy some status in human society. Human society is complicated
enough already.
Next: Remarks
Up: Humans and Robots
Previous: A conjecture about human
John McCarthy
Mon Jul 15 13:06:22 PDT 2002