490 likes | 666 Vues
Literacy Assessments. Literacy Workgroup Marcia Atwood Michele Boutwell Sue Locke-Scott Rae Lynn McCarthy. Getting Ready. Would like to have an opening activity here to activate background knowledge. Assess Frequently. In order to determine reading problems early
E N D
LiteracyAssessments Literacy Workgroup Marcia Atwood Michele Boutwell Sue Locke-Scott Rae Lynn McCarthy
Getting Ready • Would like to have an opening activity here to activate background knowledge
Assess Frequently • In order to determine reading problems early • In order to monitor which skills are developing and which skills need more explicit instruction • In order to allow teachers to make informed instructional decisions at the point of need.
Types of Reading Assessments • Standards based assessments • General outcome measures • Diagnostic • Progress monitoring
General Outcome Measures • General Outcome (Screening) Measures • A simple set of procedures that teachers can use to plan, adapt, individualize, and evaluate instructional programs for their students • (Christine A. Espin, Anne Foegen) and (Deno, 1985; Deno & L.S. Fuchs, 1987; L.S. Fuchs & Deno, 1991; Shinn, 1989) Provide a bottom-line evaluation of the effectiveness of a reading program and/or a teacher’s instruction to determine which children will need additional support in achieving important reading outcomes
Progress Monitoring • Determine if students are making adequate progress at their instructional level • Determine if need more intervention to close the achievement gap • The use of direct, repeated measurement of student progress toward long-range instructional goals • Standard tasks used as indicators of student proficiency • (Espein and Foegen, 1996)
Progress Monitoring Frequency • Too few data points taken too infrequently means that students will stay in ineffective interventions too long • As the frequency of progress monitoring increases, the probable strength of the data’s ability to reliably inform instructional increases • 2x/week after 10 weeks: excellent with 1 probe • 1x/week after 10 weeks: excellent with 1 probe • Every 3 weeks after 10 weeks: poor with median of 3 probes Pearson Education, Inc
Frequency of Assessment and Student Achievement • Bangert-Drowns, R.L. Kulik J.A. & Kulik, C.L.C, (1991), Effects of frequent cclassroom testing. Journal of Educational Research, 85. 89-99
Diagnostic Assessments • Help teachers plan instruction by providing in-depth information about students’ skills and instructional needs that impact general outcome measures
International Reading Association Standards for Literacy Assessment • Interests of the students are paramount in the assessment • The teacher is the most important agent of assessment • The primary purpose of assessment is to improve teaching and learning • Assessment must reflect and allow for critical inquiry into curriculum and instruction
Assessment must recognize and reflect the intellectually and socially complex nature of reading and writing and the important roles of school, home, and society in literacy development. • Assessment must be fair and equitable • The consequences of an assessment procedure are the first and most important consideration in establishing the validity of the assessment.
The assessment process should involve multiple perspectives and sources of data. • Assessment must be based in the local school learning community, including active and essential participation of families and community members • All stakeholders in the educational community-students, families, teachers, administrators, policy makers, and the public-must have an equal voice in the development, interpretation and reporting of assessment information.
Families must be involved as active, essential participants in the assessment process.
Before Assessing • The reason for the assessment and the use of the data must be clear • What do you want to know?
What do you do with the data? • Identify the need • Validate the need • Plan the intervention • Determine the individual expected rate of improvement • Evaluate the intervention • Review outcomes
Based on your data, you can determine… • What’s working? • What’s not working? • Who is on target for achieving standards and benchmarks? • Who is at risk for reading difficulties? • Who is not making progress adequate enough to close the gap?
Identify System Patterns • Are there components of the big ideas mastered/not mastered by the majority of students? • Are there differences in the performance of subgroups ( grades, teachers etc.)? • Are there similarities among students’ performance? • Are additional data needed?
In the classroom teachers can use data to… • Group students for instruction • Target specific reading concepts and skills that students have not mastered • Determine instructional intensity • Monitor student progress • Identify personal professional development interests and needs
Changes can be made in… • Intensity (explicit, targeted, strategic) of instruction • Group size • Amount of time in intervention • Change in program • Assessment procedures
Instructional Difference “We have research to indicate that when a student is performing below the grade level of the reading instruction being delivered in the general education program, the classroom program has little effect on the target student. Instead, tutoring accounts for the student’s growth.” Dr. Lynn Fuchs Reading Rockets
Reading Instruction Must… • Be explicit and systematic • Be paced appropriately • Be based on student assessment data • Allow opportunities to see it, guided practice, independent practice • Be based on research
In order to achieve this, teachers must…. • Understand that assessment is an important part of instruction • Understand how to administer different types of assessment and when to administer them • Analyze the data in order to use it to inform instruction
Assessing the Big 5 • Phonemic Awareness • Phonics • Vocabulary • Fluency • Comprehension
Sample Phonics Assessments Pronouncing the phonetic elements in isolation • Pronouncing a sampling of phonetic elements in sentences real and nonsense words • Really Great Reading Diagnostic Decoding Surveys (complimentary download from www.reallygreatreading.com) • DIBELS Nonsense Word Fluency • Quick Phonics Screener, Read Naturally
Importance of Vocabulary • The second most important root cause for comprehension deficits • It is rare to find a child who is good at decoding and has good vocabulary knowledge but is weak in comprehension • Vocabulary knowledge predicts word reading ability – at first grade, it predicts comprehension 10 years later
Dimensions of VocabularyMaryanne Wolf, Ph.D. • Incrementality – degrees of knowing • Multidimensionality – morphology, syntax, pragmatics • Interrelatedness – features of a word and how it relates to other words • Polysemy – knowing the multiple meanings of a word; predicts comprehension, aids in word recognition in and out of context; 1/3 of English words are polysemous
Vocabulary Assessment • No universal screening tool yet • Receptive Vocabulary – matching a picture to a word • Definitional Knowledge – describing the meaning of the word • Multiple Meanings (semantics) - identify both pictures that represent the word (Communication Intent (syntax) through ambiguous sentences and figurative language)
Receptive Vocabulary • Point to the picture that means road.
Definitional Knowledge • “Tell me what map means.” • DIBELS – Word Use Fluency (WUF) "Listen to me use the word in a sentence, (pause) "The rabbit is eating a carrot." Your turn, "rabbit.” Categories – choose the pictures that belong together
Polysemous Word Knowledge • Self-assessment • Ambiguous Sentences – find the two pictures that go with the sentence. “We need a new bat.” • Pragmatics – “What does the girl mean when she says, ‘Go fly a kite’?”
Vocabulary Connections • The extent of students’ vocabulary knowledge relates strongly to their reading comprehension and overall academic success. (Fran Lehr and associates commenting on the persistent evidence provided by Baumann, Kame’enui & Ash, 2003; Becker, 1977; Davis, 1942; Whipple, 1925)
Oral Reading Fluency • Predictor of later reading outcomes (Fuchs, Fuchs, Hosp, & Jenkins, 2001; Shinn, 1998) Richard Wagner, FCRR: • Primary concern to prevent reading difficulties is decoding; it is the most serious threat to reading achievement • Studies with thousands of children, replicated 3 times: nearly all poor comprehenders had decoding AND vocabulary deficits • Only .2% to .5% of poor comprehenders were adequate decoders
One Minute Probes • “counting the number of words read correctly from text under standardized 1-minute testing conditions is an excellent indicator of general reading achievement, including comprehension, for most students,” Advanced Applications of Curriculum –Based Measurement, Mark R. Shinn (1998)
Rate of Improvement • Fuchs et al. (1993) reasonable expectations for average, poor and disabled readers
Calculating Expected Growth • Data points fluctuate significantly • Establish an aim (goal) line based on expected growth per week and the number of weeks of instruction • Calculate the trend line • Calculate the R-squared value Shows how closely the estimated values for the trend-line corresponds to the actual data
Correlation Between ORF and Statewide Assessments • Fluency rate of third graders and the third grade end of year state assessment .66 Crawford et al. (2001) • Spring ORF and Oregon reading assessment .67 Good et al. (2001) • ORF and reading comprehension in Iowa .80 Fuchs et al. (2001) (FCRR - .50 makes us jump around the room)
Assessing Fluency • DIBELS (DIBELS Next) Oral Reading Fluency (Grades 1-6) • Aimsweb Reading CBM, Pearson, Inc. (Grades 1-8) • DRA2 Diagnostic Reading Assessment 2 • Intervention Central Oral Reading Fluency Passage Generator • Easy CBM (Grades 1-8) • Ohio Literacy Alliance (Grades 9-12)
Means to the End • Comprehension is influenced by: • Accurate and fluent word reading • Vocabulary and linguistic competence • Conceptual and factual knowledge • Knowledge and skill in the use of cognitive knowledge about what to do when comprehension breaks down • Reid Lyon, NICHD
Sample Comprehension Assessments • DIBELS DAZE • Project PROACT Maze Reading Passages • Vanderbilt University (flora.murray@vanderbilt.com) • DRA-2 – inferential and factual questions • Writing samples • DIBELS Oral Reading Retell • Easy CBM – multiple choice
Maze CBM • Passage of connected text • First sentence intact • Every nth (e.g. 7th) word deleted • 3 choices provided • Timed – 2:30 – 3:00 minutes • Fuchs and Fuchs (1992) found that for students with mild disabilities, the stability of maze data was higher than that of other reading measures.
Motivation to Read • Expectancy-Value Theory (Eccles, 1983) • Motivation is dependent on two factors: • The extent to which the person expects success or failure • The value or overall appeal that the person associates with the task
Decrease in Motivation to Read • Motivation to read decreases with age. The decline begins at or about the fourth grade. • (Durik et al., 2006; Kush & Watkins, 1996; McKenna et al., 1995)
Assessing Motivation • Interest inventories • Motivation to Read Profile (MRP; Gambrell, Palmer, Codling, & Mazzoni, 1996) • Reading Survey • Conversational Interview
Let’s Look at Data • This is literacy data for a middle school • What do the data tell you? • What questions do they raise? • What is missing?
How would you lead this district towards a QIP goal? • Using the LQI tools that you have in your packet, determine what is needed specifically in the area of assessment for literacy in this district. • What would your goal look like? • What would your system objective look like? • What would your student objective look like?
What Assessment Activities Would You Recommend? • How would you assess the goal? • What activities would you use to analyze the data?