Page 50 -
P. 50

44                                                        B. Edmonds

            data the first time the model was tested, it was then adjusted until it did work, or,
            alternatively, only those models that fitted the out-of-sample data were published (a
            publishing bias). Thus, in these cases, the models were not tested against predicting
            the out-of-sample data even though they were presented as such. Fitting known data
            is simply not a sufficient test for predictive ability.
              There are many reasons why prediction of complex social systems fails, but two
            of the most prominent are (1) it is unknown what processes are needed to be included
            in the model and (2) a lack of enough quality data of the right kinds. We will discuss
            each of these in turn.
            1. In the physical sciences, there are often well-validated micro-level models (e.g.
              fluid dynamics in the case of weather forecasting) that tell us what processes are
              potentially relevant at a coarser level and which are not. In the social sciences,
              this is not the case—we do not know what the essential processes are. Here, it is
              often the case that there are other processes that the authors have not considered
              that, if included, would completely change the results. This is due to two different
              causes: (a) we simply do not know much about how and why people behave
              in different circumstances, and (b) different limitations of intended context will
              mean that different processes are relevant.
            2. Unlike in the physical sciences, there has been a paucity of the kind of data we
              would need to check the predictive power of models. This paucity can be due
              to (a) there is not enough data (or data from enough independent instances) to
              enable the iterative checking and adapting of the models on new sets of unknown
              data each time we need to, or (b) the data is not of the right kind to do this. What
              can often happen is that one has partial sets of data that require some strong
              assumptions in order to compare against the predictions in question (e.g. the data
              might only be a proxy of what is being predicted, or you need assumptions in
              order to link sets of data). In the former case, (a), one simply has not enough to
              check the predictive power in multiple cases, so one has to suspend judgement as
              to whether the model predicts in general, until the data is available. In the latter
              case, (b), the success at prediction is relative to the assumptions made to check
              the prediction.
              A more subtle risk is that the conditions under which one can rely upon a model
            to predict well might not be clear. If this is the case, then it is hard to rely upon the
            model for prediction in a new situation, since one does not know its conditions of
            application.




            4.2.3 Mitigating Measures

            To ensure that a model does indeed predict well, one can seek to ensure the
            following:
            • That the model has been tested on several cases where it has successfully
              predicted data unknown to the modeller (at the time of prediction)
   45   46   47   48   49   50   51   52   53   54   55