Sunday, November 25, 2012

Post 13: Assessment



Blog Post 13
November 26, 2012   
Shohamy (2000) pages 201-215
Brown Chapters 23-24

            In Shohamy’s article, she discusses how discourse can affect language testing.  There are various elements of the discourse which can affect a language test. “The issue is not the validity of the discourse elements within communicative language competence, but rather the extent to which these features manifest themselves in testing situations and their effect on test takers’ scores (Shohamy 203).
 First off, I believe that the test writer should consider all of their test takers.  Their prior knowledge needs to be taken into account.  According to the article, prior knowledge could give some test takers advantages. Hock (1990) showed both language proficiency and prior knowledge predicted test takers scores on a cloze test.  Yet Clapham (1994) examined this effect of background knowledge in relation to test performance and found no such effect.  Either way, it should be considered.  Let’s say there is a reading passage about Halloween or Thanksgiving.  Two American holidays.  Some students may not have any background knowledge and others might.  This could in turn alter their test scores.
In Brown’s chapter 23, he discusses the five basic principles for designing effective tests and assessments.  The first is practicality.  A good test needs to be practical in the sense of financial limitations, time constraints, ease of administration and scoring and interpretation.
The second principle is reliability.  A reliable test is consistent and dependable.  There are many things that can cause a test to be unreliable, though.  Some of these are the construction of the test, administration of the test, the test taker, and the scoring.  In my ENG 346 class this is something that we have studied in depth.  There is person-related reliability, instrument related reliability and rater related reliability.
The third principle is validity.  This is a very complex criterion for a good test.  There are three main types of validity: content, face, and construct.  Each needs to be assessed carefully before test administration.  The fourth principle is authenticity.   This is like saying that the test you have created is a task that is likely to be enacted in the real world. According to Brown (451) there are many things that need to be in place to assure this.  For example, the language of the test needs to be as natural as possible.  Items should be contextualized rather than isolated.  Topics and situations are interesting, enjoyable, and humorous.  Some thematic organization to items is provided, such as through a story line or episode.  The tasks should represent, or closely approximate, real-world tasks.
Brown’s chapter 24 discusses classroom-based assessment.  The first topic discussed is the difference between norm-referenced and criterion-referenced tests.  Norm-referenced tests are designed to compare students by dispersing the average student scores along a bell curve.  This can determine students’ individual performance in comparison to others. While developing a norm-referenced test, the publisher must analyze the correlation between parts of the test and reliability measure.  The student is also is placed in percentage among their peers.  As for interpretation of results, a norm-referenced test’s scoring is based on the mean and standard deviation.  The assessor must also consider the standard error of measurement, which is the difference between the score the student receives and the true score, which is what the student actually knows.  Essentially, norm-referenced tests relates one student’s performance to that of other students’ performance.  It does not, however, usually permit you to identify specifically what students have or have no learned.  It does not enable you to isolate aspects of instruction that need additional attention in order to enhance student achievement.
The purpose of criterion-referenced tests is to classify people according to whether or not they are able to perform some task or set of tasks satisfactorily. The tasks are set, and the performances are evaluated. Constructing this test involves describing language objective, selecting appropriate test tasks, devising and assembling test tasks, and devise a scoring system.  Choosing the type or types of tasks to include in a test depends on a combination of factors: instructional objectives, students’ proficiency level, instructional activities, and available testing resources. Tasks can be open-ended or closed-ended. The tests are constructed such that a ‘pass’ is obtained by completing the great majority of the test tasks successfully.  Students take only the test or tests on which they are expected to be successful. This test is also useful for isolating aspects of instruction that need additional attention in order to enhance student achievement.
In the end, assessment helps everyone.  It helps the instructors identify where their student are and what needs to be taught.  They can reflect on their teaching strategies and see if they need to be altered.  Students benefit from assessments from washback, such as self-confidence gain.  Students also learn where they are struggling and can identify where they need help.

No comments:

Post a Comment