334 results on '"Shepard, Lorrie A."'
Search Results
2. Educational accountability: the use and abuse of assessment data
- Author
-
Saldaña, Christopher M., primary and Shepard, Lorrie A., additional
- Published
- 2023
- Full Text
- View/download PDF
3. What Should Psychometricians Know about the History of Testing and Testing Policy?
- Author
-
Shepard, Lorrie A.
- Subjects
- *
CAREER development , *EDUCATIONAL tests & measurements , *COMMON Core State Standards , *SPECIAL education , *INTELLIGENCE tests - Abstract
In 2023, a National Council on Measurement in Education Presidential Task Force developed a consensus framework for foundational competencies in educational measurement to guide graduate programs and subsequent professional development. This article elaborates on the social, cultural, historical, and political context subdomain from that framework. A graduate course on the history of testing and testing policy in the United States is proposed to help measurement professionals develop an understanding of historic belief systems and theories of action that affect every aspect of testing applications—definition of constructs, instrument design, respondents' interactions, interpretations and use of results, and both intended and unintended consequences. Two, accessible, key readings are proposed for each of 14 weeks addressing the following topics: IQ testing and deficit perspectives; special education placements, disproportionality, and accommodations; grade retention and tracking; college admissions testing; standards‐based reforms; 1990s performance assessment innovations; NCLB and school accountability; achievement gaps and opportunity to learn; NAEP and international assessments; standard setting and NAEP achievement levels; Common Core State Standards and ESSA; formative assessment and research on learning; culturally responsive assessment. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Classroom Assessment to Support Teaching and Learning
- Author
-
SHEPARD, LORRIE A.
- Published
- 2019
5. The Relevance of Learning Progressions for NAEP
- Author
-
American Institutes for Research, National Center for Education Statistics (ED), Shepard, Lorrie, Daro, Phil, and Stancavage, Frances B.
- Abstract
"Learning progressions" are one of the most important assessment design ideas to be introduced in the past decade. In the United States, several committees of the National Research Council (NRC) have argued for the use of learning progressions as a means to foster both deeper mastery of subject-matter content and higher level reasoning abilities. Consideration of learning progressions is especially important in the context of the new Common Core State Standards (CCSS) and Next Generation Science Standards (NGSS) that attend specifically to the sequencing of topics and skills across grades to ensure attainment of college and career expectations by the end of high school. In this paper we address the question: Should more formally developed learning progressions be considered for the future design of the National Assessment of Educational Progress (NAEP)? After a brief overview of the research on learning progressions, we describe the idealized model whereby shared, instructionally grounded learning progressions--once developed--could be used to link classroom-level assessments with large-scale assessments such as NAEP. At the same time, we also consider potential problems. In particular, learning progressions--which require agreed-upon instructional sequences--could be problematic in the context of a national assessment program intended to be curriculum neutral (i.e., not favoring one state's or district's curriculum over another). Finally, we use a sample of NAEP and Balanced Assessment in Mathematics (BAM; Mathematics Assessment Resource Service, 2002, 2003) items to explore the possibility of constructing "quasi learning progressions" that could be used to illuminate the substantive meaning of the NAEP achievement results. An appendix presents: Items in Learning Progressions. [For the main report, "Examining the Content and Context of the Common Core State Standards: A First Look at Implications for the National Assessment of Educational Progress," see ED545237.]
- Published
- 2013
6. How Middle-School Mathematics Teachers Use Interim and Benchmark Assessment Data. CRESST Report 807
- Author
-
National Center for Research on Evaluation, Standards, and Student Testing, Shepard, Lorrie A., Davidson, Kristen L., and Bowman, Richard
- Abstract
In 2001, the enactment of the No Child Left Behind Act intensified the pressure on school districts to raise test scores, close achievement gaps, and turn around low-performing schools. In response, a large number of school districts have adopted interim or benchmark assessments to be administered periodically throughout the school year in anticipation of annual state tests. This report focuses on middle-school mathematics teachers' uses of interim and benchmark assessment results. We present findings from two-stage interviews with 30 teachers in seven districts across two states. While teachers' uses of assessment information varied, few gained substantive insights about students' mathematical understanding. Instead, teachers most frequently retaught standards or items with the lowest scores and focused on procedural competence. Although many teachers expressed an interest in using assessment results to inform instruction, they reported minimal professional development to this end, and often had a different understanding regarding the intended use of the assessments than did district leaders. Examples of big picture growth and proficiency assessment information are appended. (Contains 1 table, 2 figures and 3 footnotes.)
- Published
- 2011
7. Sensitivity of NAEP to the Effects of Reform-Based Teaching and Learning in Middle School Mathematics
- Author
-
American Institutes for Research, Stancavage, Fran, Shepard, Lorrie, and McLaughlin, Don
- Abstract
This study is a validity study of the National Assessment of Educational Progress (NAEP), intended to test the adequacy of NAEP for detecting and monitoring the effects of mathematics education reform. The current study design was intended to support a comparison of the relative effectiveness of three different types of large-scale assessments--"Balanced Assessment in Mathematics" (BAM), NAEP, and state assessments--for measuring the learning gains of students participating in a well-implemented reform mathematics curriculum. To provide a context for assessing student learning where the authors could be reasonably certain of observing substantial learning gains in mathematics over the course of a school year, they selected National Science Foundation's (NSF's) Connected Mathematics Project (CMP). Although the authors had initially hypothesized that BAM, being more closely aligned with the reform curriculum, would reveal larger gains than NAEP, they found that both assessments were equally sensitive to the gains of their sample of students in CMP classrooms, and NAEP appeared better able to detect gains in the algebra classrooms. This was true even though the BAM test required twice as much time to administer as the NAEP test. Three appendices are included: (1) Sample NAEP Items; (2) Sample BAM Task; and (3) Analyses Using Booklet Percent Correct Metric. (Contains 21 tables, 1 figure and 8 footnotes.) [This report is based on work that was jointly supported by NCES (contract ED-01-CO-0026-005) and the National Science Foundation (grant 454755).]
- Published
- 2009
8. Standards, Assessments, and Accountability. Education Policy White Paper
- Author
-
National Academy of Education, Shepard, Lorrie, Hannaway, Jane, and Baker, Eva
- Abstract
Standards-based education reform has a more than 20-year history. A standards-based vision was enacted in federal law under the Clinton administration with the 1994 reauthorization of the Elementary and Secondary Education Act (ESEA) and carried forward under the Bush administration with the No Child Left Behind Act (NCLB) of 2001. In a recent survey of policy makers, standards were acknowledged as the central framework guiding state education policy. Yet, despite this apparent unanimity about the intuitively appealing idea of standards, there is great confusion about its "operational" meaning: exactly what should the standards be, how should they be set and by whom, and how should they be applied to ensure rigorous and high-quality education for American students are the central questions that challenge policy makers and educators. For example, "content" standards (subject-matter descriptions of what students should know and be able to do) are often confused with "performance" standards (which are more like passing scores on a test), and very different theories of action are used to explain how standards-based reforms are expected to work. Ambitious rhetoric has called for systemic reform and profound changes in curriculum and assessments to enable higher levels of learning. In reality, however, implementation of standards has frequently resulted in a much more familiar policy of test-based accountability, whereby test items often become crude proxies for the standards. This disconnect between rhetoric and reality is one of the reasons for the failure of prior reforms. Standards-based education is still the core idea guiding education policy and education reform. But the foregoing issues need to be addressed if the promises of standards-based education are to be kept. As yet, neither state content standards nor state tests reflect the ambitions of standards-based reform rhetoric, and the link between high expectations for all students and capacity building has been almost forgotten. The intentions of standards-based education--to focus greater attention on student learning, to ensure the participation and success of all students, and to provide guidance for educational improvement--are in the best interest of the country. This paper offers the following recommendations: (1) The federal government should encourage the redesign and clear connection of content and performance standards--and the curricula, teacher training, and high-quality assessments to go with them--with the goal of developing clearly articulated statements of the expected progression of learning. Efforts to develop these components may involve partnerships among states, universities, groups of teachers, scholars, and the private sector; (2) The federal government should support research on accountability system indicators to reflect both the status and growth of students. Performance standards should set ambitious but realistic targets for teaching and learning, and they should communicate to the public, parents, educators, and students themselves what is to be learned. Assessment results should be reported in ways that recognize progress all along the achievement continuum; (3) The federal government should support the redesign and ongoing evaluation of accountability systems to ensure that they contribute to school improvement. Less than satisfactory school performance should trigger closer investigation of school operations before remedies or sanctions are applied, and stellar performances should also be verified. Different investigative approaches, including audit assessments, data-driven analyses, or expert constituted inspectorates, should be considered; and (4) The federal government should support an intensive program of research and development to create the next generation of performance assessments explicitly linked to well-designed content standards and curricula. (Contains 77 notes.)
- Published
- 2009
9. A Survey of Teachers' Perspectives on High-Stakes Testing in Colorado: What Gets Taught, What Gets Lost. CSE Technical Report.
- Author
-
California Univ., Los Angeles. Center for the Study of Evaluation., National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Taylor, Grace, Shepard, Lorrie, Kinner, Freya, and Rosenthal, Justin
- Abstract
Using a random sample of 1,000 Colorado teachers, this study surveyed the effects of standards, the Colorado Student Assessment Program (CSAP), and school report cards on instruction and test-related practices. Findings show that standards were perceived to have greater impact on improving instruction than did testing. Teachers said they aligned their curriculum, instruction, and lessons to the Colorado standards by adding important content. Attention to the state standards improved the quality of writing instruction and focused instruction in reading, probability, geometry, and mathematics problem-solving explanations. The reported effects of CSAP testing were more mixed. Attention to CSAP improved writing instruction but shifted instruction away from social studies and science, increased the time spent on test format practice, and lowered faculty morale. Nine appendixes provide additional information, including survey responses. (Contains 12 tables and 26 references.) (Author/SLD)
- Published
- 2002
10. Teacher Effects as a Measure of Teacher Effectiveness: Construct Validity Considerations in TVAAS (Tennessee Value Added Assessment System).
- Author
-
Kupermintz, Haggai, Shepard, Lorrie, and Linn, Robert
- Abstract
This paper examines the validity of measures of teacher effectiveness from the Tennessee Value Added Assessment System (TVAAS). It considers the following claims regarding teacher effects: (1) they capture teachers' unique contributions to student learning adequately; (2) they reflect adequate standards of excellence for comparing teachers; (3) they provide useful diagnostic information to guide instructional practice; and (4) student test scores capture desired outcomes of teaching adequately. Previous analyses of the TVAAS model and simulation studies highlight potential weaknesses in the system and identify gaps in the current record of empirical evidence bearing on its validity. (Contains 2 figures and 13 references.) (Author/SLD)
- Published
- 2001
11. The Role of Classroom Assessment in Teaching and Learning. CSE Technical Report.
- Author
-
California Univ., Los Angeles. Center for the Study of Evaluation., Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Center for Research on Education, Diversity and Excellence, Santa Cruz, CA., and Shepard, Lorrie A.
- Abstract
This paper develops a framework for understanding a reformed view of assessment, in which assessment plays an integral role in teaching and learning. The proposed model is consistent with current assessment reforms being advanced across many disciplines. Three background sections of the report describe: (1) curriculum and psychological theories that have shaped methods of instruction, conceptions of subject matter, and testing methods in the past; (2) a conceptual framework based on new theories and new relationships among curriculum, learning theory, and assessment; and (3) connections between classroom uses of assessment and external accountability systems. The fourth and fifth sections elaborate the model for classroom assessment based on social-constructivist principles, arguing for the substantive reform of assessment and its use in classrooms to support learning. The final section outlines the kinds of research studies that will be needed to help realize a reformed vision of classroom assessment. (Contains 6 figures and 186 references.) (SLD)
- Published
- 2000
12. Learning Progressions as Tools for Assessment and Learning
- Author
-
Shepard, Lorrie A.
- Abstract
This article addresses the teaching and learning side of the learning progressions literature, calling out for measurement specialists the knowledge most needed when collaborating with subject-matter experts in the development of learning progressions. Learning progressions are one of the strongest instantiations of principles from "Knowing What Students Know", requiring that assessments be based on an underlying model of learning. To support student learning, "quantitative" continua must also be represented "substantively", describing in words and with examples what it looks like to improve in an area of learning. For formative purposes, in fact, "qualitative" insights are more important than "scores". By definition, learning progressions require iterative cycles of development so as to build in horizontal coherence among curriculum, instruction, and assessment. Learning progressions are also an important resource for teacher learning. With accompanying professional development and institutional supports, they can help teachers engage their students in richer and more equitable learning experiences. Examples are cited whereby learning progressions can be used to help teachers improve their skills in setting learning goals, interpreting student ideas in relation to a progression, and responding to student ideas with specific interventions that serve to move learning forward.
- Published
- 2018
- Full Text
- View/download PDF
13. Design Principles for New Systems of Assessment
- Author
-
Shepard, Lorrie A., Penuel, William R., and Davidson, Kristen L.
- Abstract
The Every Student Succeeds Act grants states new flexibility to create more balanced assessment systems with a greater role for formative assessment. Drawing on lessons learned over three decades of research and reform, we argue that state and local leaders should take the lead in designing new assessments guided by two core principles: First, make assessments coherent with rich curriculum and instruction; second, ground this integration of curriculum, instruction, and embedded assessments in equity-focused research on learning.
- Published
- 2017
- Full Text
- View/download PDF
14. Inclusion of Limited-English-Proficient Students in Rhode Island's Grade 4 Mathematics Performance Assessment. CSE Technical Report 486.
- Author
-
California Univ., Los Angeles. Center for the Study of Evaluation., National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Center for Research on Education, Diversity and Excellence, Santa Cruz, CA., Shepard, Lorrie, Taylor, Grace, and Betebenner, Damian
- Abstract
The effect of testing accommodations, such as extra time, oral reading of the assessment, or small group testing, on the participation and performance levels of limited-English-proficient students (LEP) on the Rhode Island Grade 4 Mathematics Performance Assessment was studied. A pilot study was conducted with 22 classes of students to provide preliminary evidence of the relative validity of both the performance assessment and the traditionally used Metropolitan Achievement Test (MAT) for language-minority students in comparison with general education students. The statewide data show a clear increase in the number of LED and special education students participating in the performance assessment compared to those who took the MAT. This was probably a result of the availability of accommodations or the test instructions, which stressed full inclusion. Accommodations consistently raised the relative position of LEP and special education students compared to their position on the MAT in the past. In the operational statewide assessment there was no way to evaluate the validity of the achievement gains associated with the use of accommodations, but for the most part the level of gain appeared reasonable. In the pilot sample, teachers' mathematics grades and teachers' standards-based ratings in mathematics could be used as validity criteria to evaluate the performance assessment and the MAT. Evidence supports the validity of the tests for general education students and for language-minority students, but is not as strong for the validity for LEP students. For a first effort, the inclusion of LEP students in the Rhode Island fourth-grade assessment appears to have been reasonably successful. An appendix discusses method to evaluate differential item functioning. (Contains 13 tables, 17 figures, and 19 references.) (SLD)
- Published
- 1998
15. Principles and Recommendations for Early Childhood Assessments.
- Author
-
National Education Goals Panel, Washington, DC., Shepard, Lorrie, Kagan, Sharon Lynn, and Wurtz, Emily
- Abstract
The first of the National Education Goals states that by the year 2000 all children in America will start school ready to learn. Pressed by demands for greater accountability and enhanced educational performance, states are developing standards and creating new criteria and approaches for assessing achievement. Calls to assess young children are also increasing. This booklet indicates how best to craft such assessments in light of young children's unique development, recent abuses of testing, and the legitimate demands for clear and useful information. Following a look at recent assessment issues and the current assessment climate, the booklet lists general principles that should guide both policies and practices for the assessment of young children; these principles address benefits, reliability and validity, appropriateness to age level and language, and parental role in assessment. The booklet then details important purposes of assessment for young children: (1) promoting children's learning and development; (2) identifying children for health and special services; (3) monitoring trends and evaluating programs and services; and (4) assessing academic achievement to hold individual students, teachers, and schools accountable. For each purpose, the definition, audience, technical requirements and age continuum are listed. Recommendations for policymakers are also presented for each purpose, and a chart outlining appropriate uses and technical accuracy of assessments change across the early childhood age continuum is included. The booklet then considers combinations of these purposes that have most often occurred in practice. Contains 33 references. (HTH)
- Published
- 1998
16. Measuring Achievement: What Does It Mean To Test for Robust Understanding? William H. Angoff Memorial Lecture Series.
- Author
-
Educational Testing Service, Princeton, NJ. Policy Information Center. and Shepard, Lorrie A.
- Abstract
In testing, educators have two competing intentions. They want to be fair in allowing students to demonstrate what they know without creating testing conventions and formats that let students pretend to know. Teaching-the-test literature shows that test scores can be inflated, meaning that they can go up without a generalized increase in knowledge. Students can appear to know what they really do not know, as can be illustrated by comparing results from the National Assessment of Educational Progress with those from more familiar standardized tests. Performance assessments are intended to overcome this problem. They are also intended to overcome the format effects of multiple-choice tests that have distorted instruction and allowed students to pretend to know more than they actually do. Even with performance assessments, students may rely on familiar, rote routines, and so pretend to know more than they really do. As a videotape presented as part of the discussion illustrates, asking in different ways is the way to assure that students really know what they are doing, and that their understandings generalize across contexts. Those who are concerned with test bias explore the opposite side of the coin--that students really know, but are not able to show their knowledge and abilities because of some aspect of the test. These two perspectives can be reconciled by careful and thoughtful assessment that approaches student knowledge in different contexts. (Contains 16 figures and 12 references.) (SLD)
- Published
- 1997
17. Trends in Early Childhood Assessment Policies and Practices.
- Author
-
Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Shepard, Lorrie A., Taylor, Grace A., and Kagan, Sharon L.
- Abstract
This study documented current early childhood assessment policies and practices, focusing on changes since the 1988 nationwide survey and on changes occurring in response to Goal 1 of the National Education Goals, performance assessments, and integrated services for young children. Data were collected through telephone surveys and in-depth interviews with the state-level early childhood or elementary coordinator or testing director, as well as Part H and IDEA coordinators, Title I coordinators, Head Start directors, and the Goals 1 contact person. Documentary evidence of policies or legislative mandates was also obtained. The findings, in descending order of prevalence, follow: (1) most states have made an effort to move away from readiness testing and kindergarten retention, and there is a perceived reduction though not elimination of these practices; (2) almost all state-mandated standardized testing for school accountability has been eliminated for children below grade 3; (3) some states and local districts are moving to new assessment forms in the early grades that are more supportive of instruction; (4) misuse of screening instruments for instructional purposes has decreased since 1988; (5) professional training is needed to understand and use new assessment forms; (6) preschool testing is largely driven by mandates for categorical programs; (7) a few states are collecting data to report on progress toward Goal 1; (8) parent involvement was an issue both for identifying children with special needs and because of parent demands for standardized testing; and (9) lack of collaboration and coordination among agencies serving young children continues to be a problem. (Contains 16 references.) (KB)
- Published
- 1996
18. Effects of Introducing Classroom Performance Assessments on Student Learning.
- Author
-
National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Colorado Univ., Boulder., and Shepard, Lorrie A.
- Abstract
The effects of performance assessments on student learning were examined in a year-long project to help teachers in 13 third-grade classrooms begin to use performance assessments as part of regular instruction in reading and mathematics. The essential research question was whether students learned more or developed qualitatively different understandings because performance assessments were introduced. Achievement results for the approximately 335 students were compared to the performance of third-grade students in the same schools the year before and to third-grade performance in matched control schools. Researchers worked with the teachers throughout the year to help them develop performance assessments congruent with their own instructional goals. Standardized achievement tests and some items from the Maryland State Department of Education's performance assessment were used to measure achievement. Overall, the predominant finding was one of no difference or no gains in student learning following the year-long effort to introduce performance assessment. Researchers indicated that they saw qualitative changes in performance, but it must be acknowledged that any demonstrated benefits in achievement were small and ephemeral. (Contains 10 tables, 4 figures, and 18 references.) (SLD)
- Published
- 1995
19. An Analysis of Parent Opinions and Changes in Opinions Regarding Standardized Tests, Teacher's Information, and Performance Assessments.
- Author
-
National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Shepard, Lorrie A., and Bleim, Caribeth L.
- Abstract
Parent opinions about standardized tests and performance assessments were examined systematically. Mutually exclusive but randomly equivalent stratified samples from schools participating in a study of performance assessment and control schools were used to measure change in parent opinion over time. Approximately one-third of parents (n=105) completed questionnaires at the beginning of the school year, one-third completed them at the end of the year (similar sample), and the remaining third supplied interview samples (n=33 and n=27, respectively). Results demonstrated that parents' favorable ratings of standardized national tests did not imply a preference for this type of educational assessment over other types of assessment for measuring student or school progress. Parents considered report cards, hearing from the teacher, and seeing graded samples of student work as more informative than standardized tests, and they wanted comparative information to measure their own child's progress. When parents had a chance to look at performance assessments through the year, they endorsed their use for district purposes and preferred them for classroom use. Survey data like the Gallup Poll showing widespread approval of standardized tests should not be taken to mean that parents are opposed to other forms of assessment. Appendixes contain the parent questionnaire and the interview protocol. (Contains 3 figures, 17 tables, and 9 references.) (SLD)
- Published
- 1995
20. Second Report on Case Study of the Effects of Alternative Assessment in Instruction. Student Learning and Accountability Practices. Project 3.1. Studies in Improving Classroom and Local Assessments.
- Author
-
National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Shepard, Lorrie, Shepard, Lorrie, and National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA.
- Abstract
Three papers are presented that summarize current project findings from a study of the actual effects of introducing new forms of assessment at the classroom level. All focus on aspects of performance assessment as an alternative to traditional assessments. "Effects of Introducing Classroom Performance Assessments on Student Learning" by Lorrie A. Shepard, and others, examines effects of performance assessment on the learning of third graders in 13 classrooms. "'How Does my Teacher Know What I Know?' Third Graders' Perceptions of Math, Reading, and Assessment" by Kathryn H. Davinroy, Carribeth L. Bliem, and Vicky Mayfield uses interviews with students in the classrooms of the larger study to explore student ideas and attitudes. "How 'Messing About' with Performance Assessment in Mathematics Affects What Happens in Classrooms" by Roberta J. Flexner reviews work with the teachers of the study's classes. Eighteen tables and six figures in the three papers present study findings. (Contain 77 references in all.) (SLD)
- Published
- 1994
21. Parent Opinions About Standardized Tests, Teacher's Information and Performance Assessments. A Case Study of the Effects of Alternative Assessment in Instruction, Student Learning and Accountability Practices.
- Author
-
National Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Shepard, Lorrie A., and Bliem, Carribeth L.
- Abstract
As part of a larger study of implementation of alternative assessment, a survey of parent opinions about standardized tests and performance assessments was conducted in three elementary schools. In the three participating schools, 3rd-grade teachers attended workshops on assessment development and implemented these practices in their classrooms. Samples of 69 parents from participating schools and 36 from three control schools were interviewed after completing questionnaires. Findings suggest that parents' favorable ratings of standardized national tests, supported by a Gallup Poll on the issue, do not imply a preference for such measures over other less formal sources of information for monitoring their children's progress or for judging the quality of education at their local schools. Parents tended to rely on the teacher to tell them how their child was doing relative to others, and they seldom mentioned comparison to external and national norms. Even for accountability purposes, parents preferred talking to the teacher and seeing student work. Most parents endorsed the performance assessment problems they saw, although a few expressed concern over the subjectivity of such measures. Twelve tables present survey and interview findings. An appendix presents excerpts from some parent interviews. (Contains 12 references.) (SLD)
- Published
- 1993
22. Will National Tests Improve Student Learning?
- Author
-
Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA., Colorado Univ., Boulder., and Shepard, Lorrie A.
- Abstract
Claims that national tests will improve student learning are explored, asking whether national examinations will ensure high-quality instruction and greater student learning and whether tests developed to meet urgent political deadlines will retain essential features of authentic curriculum-driven assessments. Part I presents research evidence on the negative effects of standardized testing, such as the effects of high stakes testing on scores, the curriculum, and instruction. The National Education Goals Panel's (NEGP's) version of national examinations is presented in Part II, with attention to their proposals intended to forestall the negative effects of traditional tests. Part III identifies curricular and technical problems that must be resolved before the NEGP's vision can be achieved. These include: (1) development of world class rather than lowest common denominator standards; (2) development of incorruptible performance tasks; (3) teacher training in curriculum and instruction; (4) high standards for all students without reinstitution of tracking; and (5) cost. If tests are developed before these problems are resolved, new tests are likely to have the same pernicious effects as the old. There is a 32-item list of references. (SLD)
- Published
- 1992
23. Chapter 1's Part in the Juggernaut of Standardized Testing.
- Author
-
Shepard, Lorrie A.
- Abstract
The place for standardized testing in Chapter 1 evaluation is discussed. There is substantial evidence available on the negative effects of high-stakes standardized testing, and there is a clear link between Chapter 1 requirements and the amount of testing in most school districts. Standardized testing is usually used to identify eligible students, evaluate the Chapter 1 program, and hold individual schools accountable. It is argued that each of these purposes can be served better by other means. Alternative assessments are needed for Chapter 1 use, but any such assessments must be removed from the tyranny of normal curve equivalent gains. Any system that is devised should be subjected to its own cost-benefit evaluation to determine the costs and side effects of program improvement monitoring. Five overhead projection figures used in the presentation are included. (SLD)
- Published
- 1992
24. Effects of High-Stakes Testing on Instruction.
- Author
-
Shepard, Lorrie A. and Dougherty, Katharine Cutts
- Abstract
The effects of standardized testing on instruction were studied in two school districts with high-stakes testing. The present study was part of a larger research project concerned with the effect of testing on instruction and student learning. A total of 360 teachers in grades 3, 5, and 6 in approximately 100 schools in two districts answered a questionnaire addressing test preparation/coaching practices and the effects of testing on instruction. A limitation of the study was the overall response rate of 42%, suggesting that the respondents were not necessarily representative of all teachers. Teachers reported that they felt pressured to improve test scores by the district administration and the media. Because of the importance of testing, teachers gave greater emphasis to basic skills instruction. They felt that content not tested suffered because of the focus on the standardized tests. Testing further distorted teaching because of the extensive time given to test preparation. While it was agreed that flagrant cheating was rare, practices that would clearly boost test scores, such as rephrasing questions, were considered to occur more frequently. Teachers were aware of extensive use of test results for external purposes such as comparisons of schools or districts. In open-ended questions, teachers could point to many benefits from standardized testing, but they felt that these were outweighed by the drawbacks. Fourteen tables summarize the questionnaire results. A 13-item list of references is included. (SLD)
- Published
- 1991
25. Psychometricians' Beliefs about Learning.
- Author
-
Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA. and Shepard, Lorrie A.
- Abstract
Beliefs that psychometricians hold about learning were examined through telephone interviews with directors of testing from all 50 states and with a sample of test directors from 50 selected school districts. Interpretations of what these measurement specialists believed were based on reanalyses of the primary narrative interview data. A majority of specialists operated from implicit learning theories that encourage the close alignment of tests with curriculum and judicious teaching of test content. These beliefs, associated with criterion-referenced testing, derive from behaviorist learning theory that requires the sequential mastery of constituent skills and behaviorally explicit testing of each learning step. This sequential facts-before-thinking model of learning is contraindicated by a substantial body of evidence from cognitive psychology. It is asserted that the hidden assumptions about learning should be examined precisely because they are covert. Formal debate among measurement specialists will help ensure that testing plays its desired role in the improvement of education. There is a 32-item list of references. Two appendices present two tables summarizing interview responses and nine figures illustrating models of measurement concepts. (SLD)
- Published
- 1990
26. Discussion of Part II
- Author
-
Shepard, Lorrie A., primary
- Published
- 2019
- Full Text
- View/download PDF
27. Representation, Voice, and Inclusion: AERA in the Year 2000
- Author
-
Glazer-Raymo, Judith and Shepard, Lorrie
- Published
- 2000
28. The Role of Assessment in a Learning Culture
- Author
-
Shepard, Lorrie A.
- Published
- 2000
29. Testing and Assessment for the Good of Education: Contributions of AERA Presidents, 1915-2015
- Author
-
Shepard, Lorrie A.
- Abstract
Early presidents of the American Educational Research Association were leaders in the testing movement. Their intentions were to improve education by means of testing, which included both IQ and achievement tests. Early measurement experts acknowledged in scholarly articles that IQ tests could not measure inherited ability of groups with vastly different opportunities to learn, and yet ability testing was promoted as a beneficial means for matching instruction to individual differences until the insights of the civil rights era in the 1960s. Standard achievement measures were developed importantly to allow valid comparisons across school systems and over time, but the representations of learning that were adequate 100 years ago came to have distorting effects on teaching and learning. Today's young psychometricians have opportunities to create new assessments in partnership with curriculum experts, but they should remain alert to the ways that well-intentioned assessment systems have been corrupted in the past.
- Published
- 2016
- Full Text
- View/download PDF
30. Evaluating Test Validity: Reprise and Progress
- Author
-
Shepard, Lorrie A.
- Abstract
The AERA, APA, NCME Standards define validity as "the degree to which evidence and theory support the interpretations of test scores for proposed uses of tests". A century of disagreement about validity does not mean that there has not been substantial progress. This consensus definition brings together interpretations and use so that it is one idea, not a sequence of steps. Just as test design is framed by a particular context of use, so too must validation research focus on the adequacy of tests for specific purposes. The consensus definition also carries forward major reforms in validity theory begun in the 1970s that rejected separate types of validity evidence for different types of tests, e.g. content validity for achievement tests and predictive correlations for employment tests. When the current definition refers to both "evidence and theory" the Standards are requiring not just that a test be well designed based on theory but that evidence be collected to verify that the test device is working as intended. Having taught policy-makers, citizens, and the courts to use the word validity, especially in high-stakes applications, we cannot after the fact substitute a more limited, technical definition of validity. An official definition provides clarity even for those who disagree, because it serves as a touchstone and obliges them to acknowledge when they are departing from it.
- Published
- 2016
- Full Text
- View/download PDF
31. In Memoriam: Robert L. Linn (1938-2015)
- Author
-
Carlson, James E., Shepard, Lorrie, and Baker, Eva
- Published
- 2016
32. Protecting Children from Inappropriate Practices. ERIC Digest.
- Author
-
ERIC Clearinghouse on Elementary and Early Childhood Education, Urbana, IL., Bredekamp, Sue, and Shepard, Lorrie
- Abstract
Suggestions for early childhood educators on how to provide developmentally appropriate practices and teaching methods are offered. Suggestions concern: (1) making decisions about entrance and placement; (2) evaluating a program's accomplishment of goals; (3) planning and individualizing curriculum and instruction; and (4) promoting appropriate policies. It is argued that the early childhood profession must agree on issues of developmentally appropriate practices; act to influence policy; and articulate standards for appropriate practice in the form of guidelines for curriculum content and assessment for prekindergarten through third grade classes. (RH)
- Published
- 1990
33. Formative Assessment: Caveat Emptor
- Author
-
Shepard, Lorrie A., primary
- Published
- 2017
- Full Text
- View/download PDF
34. Why Lessons Learned from the Past Require Haertel's Expanded Scope for Test Validation
- Author
-
Shepard, Lorrie A.
- Abstract
In his article, Haertel (this issue) asks a fundamental question about how use of a test is expected to cause improvements in the educational system and in learning. He also considers how test validity should be investigated and argues for a more expansive view of validity that does not stop with scoring or generalization (the more technical and familiar of Kane's [2006] four stages of interpretive arguments) or even with extrapolation. Rather, "a full consideration of validity requires that the interpretive argument be carried all the way through to the end," to the stage four questions of how test scores are actually used. The author applauds Haertel's assertion that testing effects should be the concern of measurement professionals and would argue further that very often testing consequences can be linked directly to the substantive adequacy of the test itself, especially to the extrapolation link in the interpretive argument. Research evidence about present-day tests and their distorting effects on teaching and learning has been repeated ad nauseam. Lessons learned from this literature bear repeating, however, for several reasons. First, the problems have not yet been solved. Second, past experience with both intended and unintended testing effects provides an important list of things to watch for in new validity studies. Finally, and most importantly, there are many newcomers to the testing, learning, and learning-verification enterprise who have no experience with the failure of tests to stand in as proxies for the real thing and certainly no experience with when this substitution leads to reasonable inferences and when it does not. Because of term limits, policy makers today have no idea that they are making the exact same claims that were made 20 years ago, nor do they know why the promises of test-driven accountability did not come true. Measurement is, by definition, about quantification. But what Haertel's indirect uses of tests remind measurement professionals is that testing is also about representation and signaling of what learning and achievement are taken to mean. Haertel asks that measurement experts take greater responsibility for the indirect as well as direct effects of testing--not for crazy departures from stated goals when a testing program was launched, but for those very claims that were intended. Indeed, most often testing programs begin with a promise to improve schooling. He asks that this claim and implied theory of action be made explicit and tackled formally as part of validity research. The author agrees. (Contains 1 footnote.)
- Published
- 2013
- Full Text
- View/download PDF
35. Validity for What Purpose?
- Author
-
Shepard, Lorrie A.
- Abstract
Background/Context: The evolution of validity understandings from mid-century to now has emphasized that test validity depends on test purpose--adding consequence considerations to issues of interpretation and evidentiary warrants. Purpose: To consider the tensions created by multiple purposes for assessment and sketch briefly how we got to where we are; furthermore, to address two critically important purposes: the accountability purpose versus the learning purpose for assessment. Research Design: This is an analytic, closing commentary to this special section. Conclusions: When a test is used as an educational reform, the theory of action behind the reform should be made explicit and that theory or series of claims and assumptions is what should be examined in the validity evaluation. As to the prospect of improving the teaching profession by the use of value-added methods, I believe that this is an overly ambitious use of a potentially useful statistical tool. As these systems are being implemented, we can and should conduct validity studies designed to detect plausible shortcomings and side effects as well as intended outcomes.
- Published
- 2013
36. What the Marketplace Has Brought Us: Item-by-Item Teaching with Little Instructional Insight
- Author
-
Shepard, Lorrie A.
- Abstract
School districts across the country have adopted interim and benchmark assessments in response to NCLB pressures to raise student achievement, despite the lack of a research base. Thus, it is especially important that well-conceived, empirical studies of the effects of such programs be carried out. Various theoretical frames can help understand how teams of teachers might begin to use data to reflect upon and adjust their instructional practices. At issue is whether an idealized data-based decision-making theory of action will play out in practice. Or, will teaching-the-test practices instead be exacerbated? What we learn from the empirical studies reviewed here is that positive examples, where assessment results were coherently linked to curriculum and instruction, were facilitated by highly committed principals and teacher leaders. But, these examples were rare. More frequently, interim assessments results appeared to be used, item by item, to reteach steps in problems that were missed without attending to underlying concepts or gaining diagnostic insights. (Contains 1 figure.)
- Published
- 2010
- Full Text
- View/download PDF
37. Commentary: Evaluating the Validity of Formative and Interim Assessment
- Author
-
Shepard, Lorrie A.
- Abstract
In many school districts, the pressure to raise test scores has created overnight celebrity status for formative assessment. Its powers to raise student achievement have been touted, however, without attending to the research on which these claims were based. Sociocultural learning theory provides theoretical grounding for understanding how formative assessment works to increase student learning. The articles in this special issue bring us back to underlying first principles by offering separate validity frameworks for evaluating formative assessment (Nichols, Meyers, & Burling) and newly-invented interim assessments (Perie, Marion, & Gong). The article by Heritage, Kim, Vendlinski, and Herman then offers the most important insight of all; that is, formative assessment is of little use if teachers don't know what to do when students are unable to grasp an important concept. While it is true that validity investigations are needed, I argue that the validity research that will tell us the most--about how formative assessment can be used to improve student learning--must be embedded in rich curriculum and must at the same time attempt to foster instructional practices consistent with learning research.
- Published
- 2009
- Full Text
- View/download PDF
38. Linking Formative Assessment to Scaffolding
- Author
-
Shepard, Lorrie A.
- Abstract
Formative assessment and instructional scaffolding are essentially the same thing. Formative assessment uses insights about a learner's current understandings to alter the course of instruction and thus support the development of greater competence. Scaffolding refers to supports that teachers provide the learner during problem solving--in the form of reminders, hints, and encouragement--to ensure successful completion of a task. Four strategies illustrate the strong connection between formative assessment and research on learning: eliciting prior knowledge, providing effective feedback, teaching for transfer of knowledge, and encouraging student self-assessment.
- Published
- 2005
39. Social Models of Learning and Assessment
- Author
-
Penuel, William R., primary and Shepard, Lorrie A., additional
- Published
- 2016
- Full Text
- View/download PDF
40. Assessment and Teaching
- Author
-
Penuel, William R., primary and Shepard, Lorrie A., additional
- Published
- 2016
- Full Text
- View/download PDF
41. The Hazards of High-Stakes Testing.
- Author
-
Shepard, Lorrie A.
- Abstract
Recounts historical experience with testing efforts. Recommends that in any testing program, the limitations of testing must be kept in mind in order for the tests to benefit student achievement. (DDR)
- Published
- 2003
42. Impact of Undergraduate Science Course Innovations on Learning
- Author
-
Ruiz-Primo, Maria Araceli, Briggs, Derek, Iverson, Heidi, Talbot, Robert, and Shepard, Lorrie A.
- Published
- 2011
- Full Text
- View/download PDF
43. EDITORIAL: Next-Generation Assessments
- Author
-
Shepard, Lorrie A.
- Published
- 2010
44. Comment
- Author
-
Shepard, Lorrie A
- Published
- 2002
- Full Text
- View/download PDF
45. Commentary on the National Mathematics Advisory Panel Recommendations on Assessment
- Author
-
Shepard, Lorrie A.
- Published
- 2008
- Full Text
- View/download PDF
46. The Role of Assessment in a Learning Culture
- Author
-
SHEPARD, LORRIE A.
- Published
- 2008
47. Accounting for Statistical Artifacts in Item Bias Research.
- Author
-
Shepard, Lorrie
- Abstract
Theoretically preferred item response theory (IRT) bias detection procedures were applied to both a mathematics achievement and vocabulary test. The data were from black seniors and white seniors on the High School and Beyond data files. We wished to account for statistical artifacts by conducting cross-validation or replication studies. Therefore, each analysis was repeated on randomly equivalent samples of blacks and whites (n's=1500). Furthermore, to establish a baseline for judging bias indices that might be attributable only to sampling fluctuations, bias analyses were conducted comparing randomly selected groups of whites. Also, to assess the effect of mean group differences on the appearance of bias, pseudo-ethnic groups were created, i.e., samples of whites were selected to simulate the average black-white difference. The validity and sensitivity of the IRT bias indices were supported by several findings. The pattern of between study correlations showed high consistency for parallel ethnic analyses where bias was plausibly present. Also, the indices met the discriminant validity test. Overall the sums-of-squares statistics (weighted by the inverse of the variance errors) were judged to be the best indices for quantifying item characteristic curve differences between groups. (PN)
- Published
- 1983
48. Inflated Test Score Gains: Is It Old Norms or Teaching the Test? Effects of Testing Project. Final Deliverable--March 1989.
- Author
-
Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA. and Shepard, Lorrie A.
- Abstract
It is increasingly recognized, following the lead of J. J. Cannell, that actual gains in educational achievement may be much more modest than dramatic gains reported by many state assessments and many test publishers. An overview is presented of explanations of spurious test score gains. Focus is on determining how test-curriculum alignment and teaching the test influence the meaning of scores. Findings of a survey of state testing directors are summarized, and the question of teaching the test is examined. Some frequently presented explanations refer to norms used; others refer to aspects of teaching the test. Directors of testing from 46 states (four states conduct no state testing) replied to a survey about testing. Forty states clearly had high-stakes testing. The most pervasive source of high-stakes pressure identified by respondents was media coverage. Responses indicate that test-curriculum alignment and teaching the test are distorting instruction. A possible solution is to develop new tests every year, changing the tests rather than the norms. Two tables present explanations for test score inflation and selected survey responses. (SLD)
- Published
- 1989
49. A Case Study of the Texas Teacher Test.
- Author
-
Center for Research on Evaluation, Standards, and Student Testing, Los Angeles, CA. and Shepard, Lorrie A.
- Abstract
The Texas Examination of Current Administrators and Teachers (TECAT) is a basic literacy test that was administered to 202,000 educators in Texas in March 1986. An in-depth case study was undertaken to examine the effects of the TECAT. The 2-year project involved 10 separate data-collection studies including structured interviews with key legislators and aides; interviews with educators, personnel directors, and Texas citizens; a compilation of legislative and press materials; an analysis of test results; participant observation in test-preparation workshops; a content analysis of relevant newspaper stories; a survey of study materials; and cost analyses. The political climate surrounding the administration of the test was studied. While it had not been expected that more than 5-10% of teachers would fail, Texas teachers spent a great deal of time preparing for the test, including sponsored workshops. The cost to the state was much greater than expected. Many teachers with poor skills did pass the test, but vocational and special education teachers were overrepresented among those who did not pass the test. Many teachers indicated, in response to an interview survey of 96 educators, that having to take the test was demoralizing, and that publicity about failures and the ease of the test had an adverse impact on the public perception of teachers. There are 13 tables of data and six illustrative figures. A nine-item list of references is included. Seven appendices contain supplemental information about the testing process, including the interview protocol and the teacher and personnel director phone surveys. (SLD)
- Published
- 1987
50. Reporting the Results of Statewide Assessment.
- Author
-
Shepard, Lorrie
- Abstract
Reporting the results of statewide assessment looms as a problem as more states pass from the planning to implementation phase in their assessment programs. When energies are focused on the purpose of the assessment, formulating objectives, and instrument construction, reporting takes a back seat because it happens last. There are some general principles to be followed in order to report effectively the results of a large scale assessment program. This paper begins with several recent references on how to report the results of large scale assessment programs. The remainder of this paper is intended to provide specific new thoughts for implementation of old principles. The ultimate success of state assessment programs will depend on how well assessment results are reported to their various audiences. In this paper, the most compelling recommendations for improving reporting practices are plan ahead, develop different reports for different audiences, and field test report formats to determine the language and content that are most meaningful to respective audiences. Reporting should receive the same careful attention as instrument construction with sufficient opportunity for feedback from intended users. (Author/DEP)
- Published
- 1975
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.