1 / 26

CRESST Conference, UCLA Sept 9th, 2004

Assessing Reading Skills in Young Children: The TBALL Project (Technology Based Assessment of Language and Literacy) *. CRESST Conference, UCLA Sept 9th, 2004. * Funding by NSF is gratefully acknowledged ( Grant No. 0326214). TBALL Overview. TBALL people and goals

jereni
Download Presentation

CRESST Conference, UCLA Sept 9th, 2004

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Assessing Reading Skills in Young Children:The TBALL Project(Technology Based Assessment of Language and Literacy)* CRESST Conference, UCLA Sept 9th, 2004 *Funding by NSF is gratefully acknowledged (Grant No. 0326214).

  2. TBALL Overview • TBALL people and goals • Challenges (with examples) • Approach and progress • Plans

  3. TBALL Team • Schools: • 28th street • UES • Hoover • Para los Ninos • Esperanza • EE • GSEIS • CS • EE • Linguistics • Psychology • Education

  4. TBALL Team • UCLA: Abeer Alwan (PI), Eva Baker (Co-PI), Alison Bailey, Christy Boscardin, Margaret Heritage, Dick Muntz, Carlo Zaniolo • Berkeley: David Pearson (Co-PI) • USC: Elaine Andersen, Shrikanth Narayanan (Co-PI) • Consultant: Patti Price • Students: 11 Graduate Students and 3 Undergraduates • Teachers (RETs): Six teachers from K-12 schools • Advisory Board: 5 internationally known experts in speech technology, reading, and education

  5. TBALL Specific Aims • Develop assessment system and tools • Helpful for teachers • Test mono and multi-lingual students consistently • Automatically score, analyze K-2 children • Investigate emerging literacy measures that are reliable indicators of later academic performance

  6. Why Technology-Based Assessment? • Teacher time constraints • Teacher knowledge constraints • Attractive activity for children • Assessment tailored to individual students needs • Valid, reliable information about students’ progress and needs

  7. Components of Assessment Select and present test materials Collect responses Analyze results

  8. Sample Challenges • What materials to present and how? • How to adapt speech recognition to children’s speech? • How to diagnosis discrepancies arising from • Pronunciation differences • Language exposure differences • How to detect distinct learner profiles?

  9. What Material to Present? • Many different aspects of reading skills • Phonemic Awareness • Letter-sound knowledge, Blending, Spelling • Word Recognition, Rate and Accuracy • Morphology, Syntax,Comprehension • How to diagnostically assess all aspects within the focus span of a young child?

  10. .. And how to present it?? • Children’s demonstration of language & cognitive skills is highly variable across contexts • Researchers need to be sensitive to ecological validity of procedures • How will our collection technique affect the data? • Will it disadvantage some children in the measures?

  11. Example of Presentation Differences: Hecht’s Results Percentage appropriate use of plural in each task 100 80 Percent correct use 60 40 20 Object Game Posttest Pretest

  12. Speech Recognition Challenges Significant progress in speech recognition Basic method is statistical pattern matching This task is easy in some ways, but… • Shorter vocal tract lengths, higher pitch • Significant intra- and inter-speaker variability • Significant variability • Different linguistic backgrounds • Misarticulations • Signal to noise ratio

  13. Reading Error or Pronunciation Difference? • How do we know that reading is correct? /k aw/ • A misreading of ‘car’ (saw first letter and guessed) • Or, a misarticulation/idiolect (can’t say ‘r’) • Or, possibly a dialect/accent issue (/jh eh s/ for ‘yes’) • We don’t know what the word is unless we know something about the system

  14. What marks “Hispanic accent” in English? Age of exposure, time of exposure are highly variable

  15. Learner Profiles Individual data is ‘messy’, but the ‘average’ line hides the two distinct groups of learners.

  16. The Biggest Challenge • Multidisciplinary collaboration • To solve these challenges requires • Engineering • Psychology, linguistics, psycholinguistics • Experts in reading, assessment, datamining • Starting from such different points of view • Difficult to integrate into one coherent view • Also the biggest opportunity • And probably essential

  17. Samples For You To Rate!

  18. Components * Present auditory, text, graphical stimuli * Measure decoding, comprehension skills * Score, analyze, and adapt to responses (Query-based datamining: monitor progress, compare, experiment) (Displays for teachers to combine data to help make decisions) Did this student improve? Which improved most? Which data set performs best? Who is teacher C? (Resources for teacher development)

  19. Development Process • Task specifications • Write items • Teacher review • Try with students • Design interface • Try interface • Teacher review • Displaying results

  20. Sampling Domain Core that all do, sampling of rest Focus on high frequency items Name letters Say Sound of letters Hear sound, point to letter Rhyming, blending Reading words, timed and not Naming images, timed and not Reading sentences, and pointing to image matching word • Oral Language • Decoding • Fluency • Comprehension

  21. Fall Battery (K example)

  22. Speech Recognition Approach • Speaker adaptation techniques • Pronunciation modeling • Noise robust (front end and/or back end) • Source and vocal tract parameter estimation

  23. Sorting/Analyzing Data • Database design allocates a place to put the collected data and its context, e.g., • Demographic info from parent, date, time, type of test • Data from test • Later the data can used for computations, e.g., • Words in isolation correct: 21/51 = 41% • Words in connected text: 20/36 = 55% • 75% of native speakers do better in connected text…

  24. Data from Christy Kim Boscardin Sorting Analyzing Data • Growth Mixture Modeling can reveal unobserved heterogeneity in the model Different developmental trajectories are accurately estimated Students who are most at-risk for reading problems can be identified

  25. Plans • Refine assessment content and tasks based on feedback • Address validity, utility, and impact for different language groups • Pilot studies on comprehension and reading in context • Train teachers, deploy in more classrooms each year • Further evaluate and refine the ASR system • Get feedback from teachers

  26. TBALL Specific Aims (Revisited) Integrate early Native speakers and not SR, datamining, displays Diverse measures Growth mixture modeling • Reading assessment, K-2 Helpful for teachers Test consistently Automate Diagnosis Prediction http://diana.icsl.ucla.edu/Tball/

More Related