1 / 49

Literacy Assessments

Literacy Assessments. Literacy Workgroup Marcia Atwood Michele Boutwell Sue Locke-Scott Rae Lynn McCarthy. Getting Ready. Would like to have an opening activity here to activate background knowledge. Assess Frequently. In order to determine reading problems early

conroy
Télécharger la présentation

Literacy Assessments

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. LiteracyAssessments Literacy Workgroup Marcia Atwood Michele Boutwell Sue Locke-Scott Rae Lynn McCarthy

  2. Getting Ready • Would like to have an opening activity here to activate background knowledge

  3. Assess Frequently • In order to determine reading problems early • In order to monitor which skills are developing and which skills need more explicit instruction • In order to allow teachers to make informed instructional decisions at the point of need.

  4. Types of Reading Assessments • Standards based assessments • General outcome measures • Diagnostic • Progress monitoring

  5. General Outcome Measures • General Outcome (Screening) Measures • A simple set of procedures that teachers can use to plan, adapt, individualize, and evaluate instructional programs for their students • (Christine A. Espin, Anne Foegen) and (Deno, 1985; Deno & L.S. Fuchs, 1987; L.S. Fuchs & Deno, 1991; Shinn, 1989) Provide a bottom-line evaluation of the effectiveness of a reading program and/or a teacher’s instruction to determine which children will need additional support in achieving important reading outcomes

  6. Progress Monitoring • Determine if students are making adequate progress at their instructional level • Determine if need more intervention to close the achievement gap • The use of direct, repeated measurement of student progress toward long-range instructional goals • Standard tasks used as indicators of student proficiency • (Espein and Foegen, 1996)

  7. Progress Monitoring Frequency • Too few data points taken too infrequently means that students will stay in ineffective interventions too long • As the frequency of progress monitoring increases, the probable strength of the data’s ability to reliably inform instructional increases • 2x/week after 10 weeks: excellent with 1 probe • 1x/week after 10 weeks: excellent with 1 probe • Every 3 weeks after 10 weeks: poor with median of 3 probes Pearson Education, Inc

  8. Frequency of Assessment and Student Achievement • Bangert-Drowns, R.L. Kulik J.A. & Kulik, C.L.C, (1991), Effects of frequent cclassroom testing. Journal of Educational Research, 85. 89-99

  9. Benchmark Assessment vs Progress Monitoring

  10. Diagnostic Assessments • Help teachers plan instruction by providing in-depth information about students’ skills and instructional needs that impact general outcome measures

  11. International Reading Association Standards for Literacy Assessment • Interests of the students are paramount in the assessment • The teacher is the most important agent of assessment • The primary purpose of assessment is to improve teaching and learning • Assessment must reflect and allow for critical inquiry into curriculum and instruction

  12. Assessment must recognize and reflect the intellectually and socially complex nature of reading and writing and the important roles of school, home, and society in literacy development. • Assessment must be fair and equitable • The consequences of an assessment procedure are the first and most important consideration in establishing the validity of the assessment.

  13. The assessment process should involve multiple perspectives and sources of data. • Assessment must be based in the local school learning community, including active and essential participation of families and community members • All stakeholders in the educational community-students, families, teachers, administrators, policy makers, and the public-must have an equal voice in the development, interpretation and reporting of assessment information.

  14. Families must be involved as active, essential participants in the assessment process.

  15. Before Assessing • The reason for the assessment and the use of the data must be clear • What do you want to know?

  16. What do you do with the data? • Identify the need • Validate the need • Plan the intervention • Determine the individual expected rate of improvement • Evaluate the intervention • Review outcomes

  17. Based on your data, you can determine… • What’s working? • What’s not working? • Who is on target for achieving standards and benchmarks? • Who is at risk for reading difficulties? • Who is not making progress adequate enough to close the gap?

  18. Identify System Patterns • Are there components of the big ideas mastered/not mastered by the majority of students? • Are there differences in the performance of subgroups ( grades, teachers etc.)? • Are there similarities among students’ performance? • Are additional data needed?

  19. In the classroom teachers can use data to… • Group students for instruction • Target specific reading concepts and skills that students have not mastered • Determine instructional intensity • Monitor student progress • Identify personal professional development interests and needs

  20. Changes can be made in… • Intensity (explicit, targeted, strategic) of instruction • Group size • Amount of time in intervention • Change in program • Assessment procedures

  21. Instructional Difference “We have research to indicate that when a student is performing below the grade level of the reading instruction being delivered in the general education program, the classroom program has little effect on the target student. Instead, tutoring accounts for the student’s growth.” Dr. Lynn Fuchs Reading Rockets

  22. Reading Instruction Must… • Be explicit and systematic • Be paced appropriately • Be based on student assessment data • Allow opportunities to see it, guided practice, independent practice • Be based on research

  23. In order to achieve this, teachers must…. • Understand that assessment is an important part of instruction • Understand how to administer different types of assessment and when to administer them • Analyze the data in order to use it to inform instruction

  24. Assessing the Big 5 • Phonemic Awareness • Phonics • Vocabulary • Fluency • Comprehension

  25. Sample Phonics Assessments Pronouncing the phonetic elements in isolation • Pronouncing a sampling of phonetic elements in sentences real and nonsense words • Really Great Reading Diagnostic Decoding Surveys (complimentary download from www.reallygreatreading.com) • DIBELS Nonsense Word Fluency • Quick Phonics Screener, Read Naturally

  26. Importance of Vocabulary • The second most important root cause for comprehension deficits • It is rare to find a child who is good at decoding and has good vocabulary knowledge but is weak in comprehension • Vocabulary knowledge predicts word reading ability – at first grade, it predicts comprehension 10 years later

  27. Dimensions of VocabularyMaryanne Wolf, Ph.D. • Incrementality – degrees of knowing • Multidimensionality – morphology, syntax, pragmatics • Interrelatedness – features of a word and how it relates to other words • Polysemy – knowing the multiple meanings of a word; predicts comprehension, aids in word recognition in and out of context; 1/3 of English words are polysemous

  28. Vocabulary Assessment • No universal screening tool yet • Receptive Vocabulary – matching a picture to a word • Definitional Knowledge – describing the meaning of the word • Multiple Meanings (semantics) - identify both pictures that represent the word (Communication Intent (syntax) through ambiguous sentences and figurative language)

  29. Receptive Vocabulary • Point to the picture that means road.

  30. Definitional Knowledge • “Tell me what map means.” • DIBELS – Word Use Fluency (WUF) "Listen to me use the word in a sentence, (pause) "The rabbit is eating a carrot." Your turn, "rabbit.” Categories – choose the pictures that belong together

  31. Polysemous Word Knowledge • Self-assessment • Ambiguous Sentences – find the two pictures that go with the sentence. “We need a new bat.” • Pragmatics – “What does the girl mean when she says, ‘Go fly a kite’?”

  32. Vocabulary Connections • The extent of students’ vocabulary knowledge relates strongly to their reading comprehension and overall academic success. (Fran Lehr and associates commenting on the persistent evidence provided by Baumann, Kame’enui & Ash, 2003; Becker, 1977; Davis, 1942; Whipple, 1925)

  33. Oral Reading Fluency • Predictor of later reading outcomes (Fuchs, Fuchs, Hosp, & Jenkins, 2001; Shinn, 1998) Richard Wagner, FCRR: • Primary concern to prevent reading difficulties is decoding; it is the most serious threat to reading achievement • Studies with thousands of children, replicated 3 times: nearly all poor comprehenders had decoding AND vocabulary deficits • Only .2% to .5% of poor comprehenders were adequate decoders

  34. One Minute Probes • “counting the number of words read correctly from text under standardized 1-minute testing conditions is an excellent indicator of general reading achievement, including comprehension, for most students,” Advanced Applications of Curriculum –Based Measurement, Mark R. Shinn (1998)

  35. Rate of Improvement • Fuchs et al. (1993) reasonable expectations for average, poor and disabled readers

  36. Calculating Expected Growth • Data points fluctuate significantly • Establish an aim (goal) line based on expected growth per week and the number of weeks of instruction • Calculate the trend line • Calculate the R-squared value Shows how closely the estimated values for the trend-line corresponds to the actual data

  37. Correlation Between ORF and Statewide Assessments • Fluency rate of third graders and the third grade end of year state assessment .66 Crawford et al. (2001) • Spring ORF and Oregon reading assessment .67 Good et al. (2001) • ORF and reading comprehension in Iowa .80 Fuchs et al. (2001) (FCRR - .50 makes us jump around the room)

  38. Assessing Fluency • DIBELS (DIBELS Next) Oral Reading Fluency (Grades 1-6) • Aimsweb Reading CBM, Pearson, Inc. (Grades 1-8) • DRA2 Diagnostic Reading Assessment 2 • Intervention Central Oral Reading Fluency Passage Generator • Easy CBM (Grades 1-8) • Ohio Literacy Alliance (Grades 9-12)

  39. Means to the End • Comprehension is influenced by: • Accurate and fluent word reading • Vocabulary and linguistic competence • Conceptual and factual knowledge • Knowledge and skill in the use of cognitive knowledge about what to do when comprehension breaks down • Reid Lyon, NICHD

  40. Sample Comprehension Assessments • DIBELS DAZE • Project PROACT Maze Reading Passages • Vanderbilt University (flora.murray@vanderbilt.com) • DRA-2 – inferential and factual questions • Writing samples • DIBELS Oral Reading Retell • Easy CBM – multiple choice

  41. Maze CBM • Passage of connected text • First sentence intact • Every nth (e.g. 7th) word deleted • 3 choices provided • Timed – 2:30 – 3:00 minutes • Fuchs and Fuchs (1992) found that for students with mild disabilities, the stability of maze data was higher than that of other reading measures.

  42. Motivation to Read • Expectancy-Value Theory (Eccles, 1983) • Motivation is dependent on two factors: • The extent to which the person expects success or failure • The value or overall appeal that the person associates with the task

  43. Decrease in Motivation to Read • Motivation to read decreases with age. The decline begins at or about the fourth grade. • (Durik et al., 2006; Kush & Watkins, 1996; McKenna et al., 1995)

  44. Assessing Motivation • Interest inventories • Motivation to Read Profile (MRP; Gambrell, Palmer, Codling, & Mazzoni, 1996) • Reading Survey • Conversational Interview

  45. Let’s Look at Data • This is literacy data for a middle school • What do the data tell you? • What questions do they raise? • What is missing?

  46. How would you lead this district towards a QIP goal? • Using the LQI tools that you have in your packet, determine what is needed specifically in the area of assessment for literacy in this district. • What would your goal look like? • What would your system objective look like? • What would your student objective look like?

  47. What Assessment Activities Would You Recommend? • How would you assess the goal? • What activities would you use to analyze the data?

More Related