Page 24 -
P. 24
Contents xvii
1 1.2 Evaluation paradigms and techniques 340
1 1.2.1 Evaluation paradigms 341
1 1.2.2 Techniques 345
1 1.3 D E C I D E: A framework to guide evaluation 348
1 1.3.1 Determine the goals 348
1 1.3.2 Explore the questions 349
1 1.3.3 Choose the evaluation paradigm and techniques 349
1 1.3.4 identify the practical issues 350
1 1.3.5 Decide how to deal with the ethical issues 351
1 1.3.6 Evaluate, interpret, and present the data 355
1 1.4 pilot studies 356
Chapter 12 Observing users 359
1 2.1 Introduction 359
12.2 Goals, questions and paradigms 360
12.2.1 What and when to observe 361
1 2.2.2 Approaches to observation 363
1 2.3 How to observe 364
12.3.1 In controlled environments 365
1 2.3.2 In the field 368
12.3.3 Participant observation and ethnography 370
12.4 Data collection 373
12.4.1 Notes plus still camera 374
12.4.2 Audio recording plus still camera 374
12.4.3 Video 374
1 2.5 Indirect observation: tracking users' activities 377
12.5.1 Diaries 377
12.5.2 Interaction logging 377
12.6 Analyzing, interpreting and presenting data 379
12.6.1 Qualitative analysis to tell a story 380
1 2.6.2 Qualitative analysis for categorization 381
12.6.3 Quantitative data analysis 384
12.6.4 Feeding the findings back into design 384
Interview with Sara Bb 387
Chapter 13 Asking users and experts 389
1 3.1 introduction 389
1 3.2 Aking users: interviews 390
13.2.1 Developing questions and planning an interview 390