1 / 15

PRINCIPLES OF LANGUAGE ASSESSMENT

PRINCIPLES OF LANGUAGE ASSESSMENT. Riko Arfiyantama Ratnawati Olivia. Job Description . Speaker I - Practicality - Reliability - Validity Speaker II - Authenticity - Washback Speaker III - Applying principles to the evaluation of classroom tests.

glenda
Download Presentation

PRINCIPLES OF LANGUAGE ASSESSMENT

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. PRINCIPLES OF LANGUAGE ASSESSMENT RikoArfiyantama Ratnawati Olivia

  2. Job Description • Speaker I - Practicality - Reliability - Validity • Speaker II - Authenticity - Washback Speaker III - Applying principles to the evaluation of classroom tests

  3. How do you know if a test is effective? 1. Practicality 2. Reliability 3. Validity 4. Authenticity 5. Washback

  4. Practically An effective test is Practical: • Is not excessively expensive, • Stays within appropriate time constraints, • Is relatively easy to administer, and • Has a scoring/evaluation procedure that is specific and time-efficient.

  5. RELIABILITY • A reliable test is consistent and dependable. If you give the same test to the same student or matched students on two different occasion, the test should yield similar results. Test II Test I First occasion Second occasion

  6. The possibilities of Reliability The fluctuations in: The students Scoring Test administration The test itself

  7. Student-Related Reliability The fluctuation in the student can be caused by the following factors: • Temporary illness, • Fatigue • A “bad day” • Anxiety • Other physical and psychological factors

  8. Rater Reliability The fluctuation in scoring can be caused by the following factors: • Human error (teacher’s fatigue) • Subjectivity • Bias (good or bad students) • Lack of attention to scoring criteria • Inexperience • Inattention

  9. Test Administration Reliability The fluctuation in administration can be caused by the following factors: • The condition (place) of the test administration e.g. listening test becomes unclear because of the street noise. • Photocopying variations • The amount of light in different parts of the room. • Variations in temperature. • The condition of desks and chairs.

  10. Test Reliability The fluctuation in the test itself can be caused by the following factors: • Time limitation in a test • The test is administered too long so the test-takers may become fatigue.

  11. VALIDITY • The extent to which inferences made from assessment results are appropriate, meaningful, and useful in terms of the purpose of the assessment. (Gronlund, 1998: 226) For example: • A valid test of reading ability actually measures reading ability. • A valid test of writing ability actually measures writing ability not grammar.

  12. Content-Related Evidence • The validity of the test depends on the content and the relation between the purpose of the test (content) and the way the test is administered (related) . For example: • To get a valid speaking test, the students should do the direct test by giving the students’ chance to perform their ability in speaking, not by giving them paper-and-pencil test.

  13. Criterion-Related Evidence Criterion-related Evidence usually falls into one of two categories: • Concurrent Validity: a test has concurrent validity if its results are supported by other concurrent performance beyond the assessment it self. E.g. a high score of the final exam of a foreign language course will be sustained by actual proficiency in the language. • Predictive Validity: the predictive validity of an assessment becomes important in the case of placement tests, admissions assessment batteries, etc. The assessment criterion in such cases is not to measure concurrent ability but to assess (and predict) a test-taker’s likelihood of future success.

  14. Construct-Related Evidence • A construct is any theory, hypothesis, or model that attempts to explain observed phenomena in our universe of perceptions. For examples: linguistic construct covers “proficiency” and “communicative competence”, and psychological construct covers “self-esteem” and “motivation”.

  15. Consequential Validity • Consequential Validity encompasses all the consequences of a test, including such considerations as its accuracy in measuring intended criteria, its impact on the preparation of test-takers, its effect on the learner, and the (intended and unintended) social consequences of a test’s interpretation and use. McNamara (2000: 54) cautions against test results that may reflect socioeconomic conditions such as opportunities for coaching that are “differentially available to the students being assessed (for example, because only some families can afford coaching)”

More Related