Page 128 - Composition in Convergence The Impact of the New Media on Writing Assessment
P. 128

VALIDITY AMD RELIABILITY       95

         numbers of students to satisfy an explanation of what is happening
         in the writing  classroom  and the  qualifier to prefer  smaller  num-
        bers of students to provide a general pattern  of activities in a com-
        position class. What  is significant  about this point is that  writing
         departments are now filled with readers from both camps and each
         side reads  student  papers. Although  assessment  indeed drives in-
         struction,  recognizing that assessment can be defined through  two
         opposite theoretical approaches illustrates the fissures that can oc-
         cur in departments regarding written  evaluation as well as the rifts
         that emerge in pedagogy.
           The schism  that  seems to  exist  in writing assessment  practices
         from  program  to program,  perhaps  even from instructor  to  in-
         structor,  mirrors  Composition's  mistrust  of  assessment.  Few
        practitioners and scholars understand tests and measurements, as
         Brian Huot (2002) rightly noted. Even fewer recognize that Com-
        position  is not  bound  to  quantitative  definitions of concepts like
        validity and reliability to describe what occurs in the classroom. It
        is possible to reconfigure these terms to accommodate the flexibil-
        ity necessary to discuss the writing  process in networked  writing
        classes  and  to  do  so without  causing  great  conflict  with  social
        constructivist  pedagogical models.  The way  to  realign  validity
        and reliability in writing  assessment in these computer-enhanced
        contexts is to think  even more qualitatively  about  evaluation.
           University of Michigan education professor Pamela A. Moss, in
        her  research on  accountable  assessment with  portfolios  (1992),
        indicated that frequently a student  writer's  growth  is made evi-
        dent  by  examining  the  qualitative aspects  of the writing. For in-
        stance,  having  instructors  look  at  the  increasing  levels  of
        complexity in student problem solving often reveals that there is a
        loss of control in mechanics or organization as student writers de-
        velop richer interpretations  of a text.  According to  Moss (1992),
        there  are  other  subtle indicators  of growth  like quality  of voice
        and  elaboration.  These characteristics  tend  to  be understated  in
        psychometric   approaches  and  subsumed  under  broader  criteria,
        which causes faculty to miss or to misinterpret  critical moments
        of a writer's  development.
           In thinking qualitatively about writing assessment in networked
        environments,  instructors  are  asked to  consider depth  instead of
        comprehensiveness   in evaluation.  This means that instead  of col-
        lecting webfolios or electronically generated assignments from  ev-
   123   124   125   126   127   128   129   130   131   132   133