Page 146 - Composition in Convergence The Impact of the New Media on Writing Assessment
P. 146

VALIDITY AND  RELIABILITY     113

        iar with electronic communication  can distinguish acceptable work in
        ways that break from  the linear holistic scoring guides, yet still retain
        the  sense of reliability that many  test-and-measurement  people want
        to  see in outcomes assessment.
           This "new" reliability does not depend on the consistency of writ-
        ing  specialists  guessing  the  same  score  to  keep  consensus  and
        interrater  reliability or to  ensure the  reliability of test  instrument,
        two situations  that frequently lead to a Panopticon of sorts in assess-
        ment settings. Instead, this new reliability insists on faculty review-
        ers who are experienced with the currency  of technological  conver-
        gence and  student  e-texts  to make decisions about  the adequacy of
        students' writing in these genres.
           Assessing for adequacy moves deep assessment closer to validity
        because  the  evaluators  have  the  opportunity  to  examine  a  fuller
        scope of the students'  writing  activities and contexts. Not only will
        the  archived data  contain  numerous  examples of writing produced
        under  various  conditions  and  for various  audiences, the  students'
        own   analytical  examinations  of  their  work  and  the  instructors'
        points  of  intervention;  the  data  should  also  reflect  the  teacher's
        comments. All these elements provide the breadth needed to make a
        valid  writing  assessment.  Deciding  whether  a  student's  archived
        writing  is acceptable depending on local criteria should pass the test
        for  face validity because the data are evaluated by local experts using
        familiar  criteria to measure the writing. Additionally, when a panel
        or team  of teachers who  are experienced in  electronic  communica-
        tion  evaluates the students'  adequacy as writers of online material,
        there  is also  predictive  validity.  That  suggests  that  evaluators  ac-
        knowledge the  students  are  reasonably  able to  do the  work  again
        later  based on examining  the archived  materials.  Concurrent  valid-
        ity  can be included in this type of evaluation  if the assessment team
        wants to measure  the students' electronic writing against students'
        F2F writing; this approach might be an especially useful step in writ-
        ing  programs  where  there  are  computer-only  sections  (and,  con-
        versely,  F2F-only sections). Testing  for  concurrent  validity  will be
        useful only if the criteria used to measure the writing remain identi-
        cal for both  sections. It is also important  to  note that in programs
        where computer-enhanced composition is under fire or where there
        is  great  skepticism,  testing  for  concurrent  validity  may  answer
        administrators'  or faculty members' concerns about  the benefits of
        using computer  technology  in the writing  classroom.
   141   142   143   144   145   146   147   148   149   150   151