Page 146 -
P. 146

8 The Importance of Ontological Structure: Why Validation by ‘Fit-to-Data’...  143


            8.1.1 Introduction to Neural Networks

            In this section, we will briefly explain what neural networks are, the mathematical
            formulas that underpin them (in Appendix 1) and the way they are structured.
            The main points we wish to introduce are that, though neural networks have
            tremendous potential to approximate data, there is nothing about their structure
            or the mathematics underpinning their functioning that necessarily reflects any
            structure or mathematics in whatever system the data were taken from.
              Neural networks were originally conceived as simulations of the brains but
            are essentially networks of nonlinear functions with parameters that are adjusted
            according to a learning rule. There are several different kinds of neural network
            mathematically speaking, and for each kind, there can be several different learning
            rules and minor adaptations and variations thereof. Biologically, a neuron is a cell
            with axons connecting it to other neurons. In an agent-based simulation of a brain,
            we would simulate a neuron as an agent and an axon as a link. The behaviour of
            the neuron is simply to emit an electrical pulse periodically. The more frequent
            the pulse, the more ‘excited’ the neuron. Connections between neurons can be
            excitatory or inhibitory. An excitatory connection means that there is a positive
            relationship between the excitation of the two connected neurons: all other things
            being equal, one neuron’s excitement increases that of the other. An inhibitory
            connection means that the relationship is negative – one neuron’s excitement
            decreases that of the other. The connection has a strength – the stronger the
            connection between one neuron and another, the more significant the relationship
            is in comparison with other neurons the neuron is connected to.
              When simulating neurons, the pulsation is ignored and the frequency of pulsation
            modelled as a variable. Simulated neurons are typically called nodes. The axons
            form the links in a directed graph connecting the nodes, and the directedness means
            that nodes have input axons and output axons. Simulated axons are typically called
            weights, largely because it is the value of the weight (representing the strength
            of the connection) that is of primary interest. The weights of a neural network
            are its parameters, and the job of the learning algorithm is to determine their
            values. The qualitative description of the behaviour of neurons is of course given
            a precise mathematical specification in simulated neural networks; this is provided
            in Appendix 1 for the benefit of those who are interested.
              A further simplification of the structure of the network is to arrange the nodes
            into distinct layers. (It can be proved that this does not lead to loss of potential
            functionality.) This simplification means that the choice of network structure is
            simply a question of determining the number of layers, and for the layers that are
            not input or output layers (the so-called hidden layers), the number of nodes to
            use in each layer. The number of nodes in the input and output layers is of course
            determined by the dimensionalities of the domain and range of the function to be
            approximated. It has been proved (Cybenko 1989; Funahashi 1989; Hornik et al.
            1989) that one hidden layer is sufficient to approximate any function. Although
            having more hidden layers can mean that the contribution of the weights closer to
   141   142   143   144   145   146   147   148   149   150   151