| 
       | 
       
        Next Previous Table of content3. Strange brains 
      
      Since there are computers, there are also attempts to 
      simulate processes of human thinking on them. The ability of an electronic 
      machine to read, manipulate and output  information in high speed, tempted 
      researchers from the outset to speculations over the equivalence of 
      computers and brains. Such speculations soon found entrance into 
      psychology. In particular the possibility of calculating machines to 
      process information parallely corresponded to approaches in  psychology 
      which regarded the brain predominantly as a 
      parallel processing system. Computers were regarded before this background as a 
      possibility of clearing up still unexplored phenomena of the human mind 
      through modelling. A good example is the Pandemonium model of 
      Oliver Selfridge (Selfridge, 1959). Pandemonium is a model for 
      visual pattern recognition. It consists of a multiplicity of  parallely 
      working demons, of which everyone is 
      specialized in recognizing a certain visual stimulus, for example a 
      horizontal bar in the center of the presented stimulus or a curvature in 
      the right upper corner. If a demon recognizes "his" stimulus, he calls  out to the 
      central master demon. This exclamation is the 
      louder, the more highly the demon estimates the probability of  a correct 
      identification. All demons work independently from one another; none is 
      affected by his neighbours. On the basis of the received information, the master demon then decides which pattern 
      constitutes the stimulus. In an advancement of the model, the demons were 
      organized  
      hierarchically in order to relieve the master demon. Between these specialized demons in Selfridge's model 
      and the actually existing feature detector 
      cells in the visual cortex exists an astonishing similarity. And 
      indeed it was the model of Selfridge and the its assumptions over 
      perception processes, which suggested for the first time that such feature detectors could exist in humans. The 
      model was in this case the reason for neurophysiologists to look for 
      appropriate cells. Thus Pandemonium is a good example of how 
      computer models can advance psychological research. On the other hand, 
      it  
      should not be forgotten that a system such as Pandemonium 
      is unable to really "see". And this is a key point for critics who grant a 
      reduced heuristic use to computer modelling, but otherwise deny any 
      equivalence between humans and machines. This is also one of the fundamental questions for the 
      development of emotional computers: the equivalence of the systems 
      "humans" and "computer". Both the AI research and the past approaches for 
      the development of emotional systems assume without further doubt that 
      "intelligence" and "emotion" i na computer are not fundamentally different 
      from intelligence and emotion in humans. This assumption abstracts to a large extent from the 
      specific hardware; "emotion" is understood as a pure software 
      implementation. But it is quite questionable whether the two systems obey 
      the same laws of development. A computer is a discrete system, which knows nothing 
      else than only two different states. By the combination of several of such 
      elements, "intermediate states" can be created; this however only on a 
      certain level of abstraction from the underlying hardware. In contrast, the physiology of the emotional and 
      cognitive system in humans by no means represents a comparable mechanism, 
      but consists, even on the lowest level, of a multiplicity of mechanisms, 
      some of which work more according to digital principles, others more 
      according to analog principles. Even one of the best researched mechanisms, the 
      function of the neurons, is not exclusively an On/Off mechanism, but 
      consists of a multiplicity of differentiated partial mechanisms - and this 
      on the level of the hardware. The simulation of such mechanisms with computers is 
      at present only possible as software. Simple neural switching patterns 
      can, up to a certain point, also be modelled by parallel computers; such a 
      modelling is possible, however, only within certain limits and ignores 
      completely chemical processes which play an important part in the human 
      brain. Picard (1997) tries to solve the problem by 
      abstracting from the difference between hardware and software and defining 
      both as "computers". She justifies this position with the argument that 
      emotional software agents can exist in an "emotion-free" hardware. A similar discussion is deals with the comparability 
      of emotions of humans and animals (see Dawkins, 1993). Here at least we 
      have a hardware of identical elements, although of different complexity. 
      In this case it is, too, not considered scientifically decided whether an 
      emotion like "mourning" is identical in humans and animals.  The affair is made more difficult still by the 
      question whether a computer can be considered in principle as a form of 
      life. In the "Artificial Life" discussion of the last years, some 
      attention was given to this question. Thus the evolutionary biologist 
      Richard Dawkins (Dawkins, 1988) holds the opinion that the ability for 
      reproduction would already be sufficient to speak of life in the 
      biological sense. Others extend the definition by the components "self 
      organization" and "autonomy". If one ignores the ethical and philosophical 
      discussion of "life" and concentrates on the aspects "self organization" 
      and "autonomy", then it is quite realistic to attribute to computers 
      and/or software these characteristics. Self organization in the sense of 
      adaptation can be observed, for example, in neural nets working with 
      genetic algorithms (see e.g. Holland, 1998). Autonomy in the reduced sense 
      can be observed in robots and/or partly in autonomously operating 
      programs, for example agents for the internet. Such programs possess also 
      the ability for reproduction, which would fulfil the third condition. The emphasis of the AI and AL research lies at 
      present on the advancement of such autonomous, self-organizing systems. 
      The models used are partly based on functional models of the human brain; 
      this should not tempt one, however, to rashly equate their operating mode 
      with that of the human brain. Especially with optimization processes of software by 
      genetic algorithms it is frequently not known to human observers which 
      self organization processes the software uses, in order to reach the 
      optimization goal. Sometimes, though, it might be useful to attribute to 
      a computer mental abilities. Thus John McCarthy, one of the pioneers of 
      artificial intelligence, explains: 
 
 The attribution of mental abilities according to 
      these remarks possesses only a functional nature: It serves to express 
      information about the state of a machine at a given time which otherwise 
      could only be expressed through lengthy and complex descriptions of 
      details.  McCarthy lists a number of mental characteristics 
      which can be attributed to computers: Introspection and self knowledge, 
      consciousness and self-consciousness, language and thinking, intentions, 
      free will, understanding and creativity. At the same time he warns to 
      equate such attributed mental qualities with human characteristics: 
 
 We now know that the last sentence must not 
      necessarily be correct. There are first examples of self-organizing and 
      self-optimizing hardware (Harvey and Thompson, 1997), whose modes of 
      functioning are not known to its human designers. And the current 
      approaches in the design of emotional computers go far beyond modelling 
      but try to develop computers whose mental qualities are not pre-defined by 
      the designer, but develop independently. Although naturally certain basic assumptions of the 
      designers flow into such systems, this approach is nevertheless 
      fundamentally different from the classical modelling approach which can be 
      observed in cognitive science. The question remains, however, whether the 
      processes in a computer which, due to this procedure, one day actually 
      develops mental qualities are identical with the processes in the human 
      body and brain. Critics of such an approach  point out the 
      fact that emotions are not comparable to purely cognitive processes, since 
      they are affected by additional factors (e.g. hormones) and also require a 
      subject. The modelling of these processes within a computer, a purely 
      cognitive construction, would therefore be impossible; the more so because 
      the subjective element is missing in a machine without which an emotion, 
      whose substantial component is a feeling, could not be felt. To this 
      argument there are several answers. On the one hand, one cannot possibly rule out that a 
      computer can possess "feelings". From an evolutionary viewpoint, computers 
      are an extremely young phenomenon which have, in their short existence, 
      made a number of giant steps. Today, there exist machines with hundreds of 
      parallel processors; impressive research progress is made with biological 
      and quantum computers. It might be just a question of time until a 
      computer does posses a similar complex hardware as the human brain. With 
      increasing complexity the probability increases, too, that such a system 
      will organize itself on a higher level. What must be laborously  programmed as 
      a "monitoring instance" today might develop into something which one day 
      might be called the "ego" of a computer. On the other hand, it would be extremely 
      anthropocentric to deny emotions to an intelligent system just because it 
      does not possess human hormones. A computer consists of a multitude of 
      "physiological" processes which could be perceived as "bodily feelings" 
      once the system has been equipped with a propioceptive subsystem. If, in 
      addition, this computer is able to learn and move, one could imagine it 
      reacting to certain situations with a change of such processes which for 
      it possess the same value as physiological changes in our body. An emotional computer must not experience emotions 
      like a human - not more than a visitor from Zeta Epsilon. Nevertheless, 
      its emotions can be as genuine to it as ours to us - and influence its 
      thoughts and actions just as much.  We can't therefore assume a priori that "emotions" 
      which are developed by a computer are comparable to human emotions. But it 
      is thoroughly justified to assume that the emotions of a computer serve 
      the same functions for it as for us humans. If this is the case, the 
      computer modelling of emotions would not only be a way to learn more about 
      human emotions; it would at the same time lay the foundations for a time 
      when intelligent systems of different "building blocks" will co-operate 
      with one another.  Next Previous Table of content | 
    
       |