|
Volume 29, No. 4, March/April, 2000. Clear answers to common-sense questions about accountability By Douglas B Reeves |
|||||
|
George Orwell (1949) helped generations of students
learn the meaning of the term "double-speak." Few practitioners of the art of double-speak have studied the concept as assiduously or applied it with as much vigor as attackers and defenders of public education. Both the strident tone of the debate between these two camps and the obscurity of the arguments employed by the debaters cause many observers to throw up their hands in dismay. Parents and policy-makers have a more direct agenda. In terms of educational accountability, they express the issues very clearly: "How's my kid doing?" "Are the schools succeeding or failing?" "What works best to help students learn?" "Do test scores prove the effectiveness of educational programs?" These are serious questions, and effective accountability systems must address them with clarity and candor. An accountability system that fails to address these common-sense questions does not deserve the support and confidence of citizens or policy-makers. Simple questions, however, do not necessarily lead to simple (or, more appropriately, simplistic) answers. So it is with educational accountability. These simple questions require complex responses, which I will explore in this article. "How's my kid doing?" A question about the performance of a specific child implies that an effective accountability system will be based on information about individual students rather than groups or classes of students. In other words, an accountability system that can supply information about the progress of a specific child must be based on a series of individual student achievement records. If these records are flawed, then the entire house of accountability is built upon a shaky foundation. One of the most important developments in educational accountability has been the "value-added" methodology developed by Professor William Sanders and his colleagues at the University of Tennessee (1998). The core of the system is a "student-to-student" comparison. While such a comparison makes common sense, it is rare. By far the majority of state and local accountability systems compare one year's class of students to the previous year's class of students - a comparison that involves almost entirely different individual students. Such group comparisons never address the fundamental question, "how is my kid doing?" Rather, parents and teachers are given the curious information that their eighth-grade students are better or worse than last year's eighth-grade students. Group comparisons tell us nothing about the progress and educational needs of individual students. This has serious implications for accountability systems that claim to evaluate educational quality, but ignore information about classroom teaching and learning. An accountability system that contains test scores alone, without the context of additional accountability information about teaching practices and curricula, is incomplete. A school district that bases its accountability system on test scores alone is analogous to a physician who evaluates physical health based upon indicators such as body temperature or blood pressure, but ignores the other medical indicators that any reasonable physician would regard as essential to a competent diagnosis. In the most bizarre cases, accountability systems not only fail to evaluate substantial parts of the school curriculum, but actually encourage behaviors contrary to those endorsed by the designers of the accountability system. Consider the case of student writing. Most states, with the resounding approval of boards, legislators and the public, have established standards for writing. These standards typically require students to express themselves coherently in expository and persuasive writing. Students must also use standard conventions of English grammar, punctuation and spelling. In order to do this, students must use the writing process (including outlines, rough drafts and editing) and draft multiple revisions in order to incorporate teacher feedback and make their final written product worthy of the demands for rigor established by state writing standards. It is axiomatic that standards implementation would require a test congruent with those standards. Simple logic and fairness demand such congruence. Unfortunately, the same logic and fairness seem to elude those legislators and education officials who establish writing standards, but endorse tests incompatible with the standards they have been designed to measure. The standard requires teachers and students to use the writing process; the test uses a single writing prompt followed by a single response. The standard requires teachers to focus on the value of effective written communication; the test requires students to correct spelling and grammar errors in a passage and complete a multiple-choice question -- without writing a single sentence. If parents or stakeholders want an honest and accurate response to the question of "How's my kid doing?" then school systems and districts must use tests or other assessments that will yield the information needed to answer this question. In other words, children need to be assessed to make sure they meet the standard, and only this information will determine how they're "doing." The most virulent critic of public education would not attend an athletic event and, desiring to know the outcome, settle for a description of the weather and field conditions. "What the heck happened during the GAME?" the critic would demand. In the context of the classroom, parents and students must learn about the process and results of teaching and learning. If the education "game" is to be taken at least as seriously as many people take their weekend athletic events, then it is reasonable to expect that the data used to evaluate the result should be related to the issue at hand: student and school performance. Thus, if we want to know how well students write, then we must ask them to write. If we want to know whether they can use the scientific method, then we should ask them to design an experiment and draw inferences from a set of data. If we wish to know whether students understand mathematics, we should ask them to use mathematics to solve real problems. "Are the schools succeeding or failing?" The second question raised in the name of common sense is built upon the first. Schools succeed only where students succeed. Thus, assessments of schools, such as accreditation or typical accountability reports, are only as effective as their evaluation of students. Accountability systems that look only at process and effort will reward a fixation on meetings, plans and strategies while ignoring results. Every initiative, including those that I have advocated -- high standards, effective assessments and rigorous accountability -- is only effective when it is built upon a foundation that soundly evaluates student achievement. Mike Schmoker (1999, 1999a) is a leading advocate of the focus on results. He has endured heaps of abuse from those who find process a perfectly suitable substitute for student achievement. While a focus on results is important, the "results" that effective accountability systems must consider extend far beyond test scores alone. This is not merely a debate over "process vs. results." In fact, a comprehensive accountability system must focus both on measurable elements of the process (specific instructional, assessment and engagement strategies), and on results (indicators of student achievement). Only with such a comprehensive view can we gain some insight into what the adults in the system can do to influence results achieved by the students. Although the evaluation of students is the foundation of a sound accountability system, an effective accountability system must base its examination of student achievement on more than test scores. Indeed, the fallacy of reporting school-wide success or failure based on single test scores has been widely documented (Bracey 1999). Teaching and learning are multiple interactive processes, the results of which are much too complex to be captured by a single score. Accountability systems that depend solely on test scores offer predetermined results: students who are capable test takers will make a school look like a success; students who are not capable test takers will give their school the label of failure. In neither case do the organization, leadership, teaching and educational practices of the school receive a meaningful evaluation. Instead, effective evaluation of the success of a school can only be measured with multiple information sources over an extended period of time. There must be several indicators that measure the performance not only of students, but also of the adult decision-makers. Such an analysis includes a consideration of resources, teaching methods and student support. Without a consideration of all of these factors, we are left with the ludicrous notion of evaluating the performance of an entire institution based on the test scores of one group of eight-year-olds during one stress-filled, springtime afternoon. Effective and comprehensive accountability systems distinguish between student achievement and school achievement, without losing sight of the fact that the latter is integrally related to the former. For example, state and district test scores might provide some indicators of student achievement. But these results are only meaningful indicators of school achievement when they are placed in the context of the specific educational strategies used by the schools. The structure of such a comprehensive system acknowledges both the distinction and the importance of both questions. "What works best to help students learn?" The third common-sense question addresses the heart of an effective accountability system. The question of program effectiveness is far more complex than the recitation of test scores representative of most accountability systems. Such a "box score" approach is the educational equivalent of giving the ranking of teams at the end of the season without shedding any light on the strategies that led teams to succeed or fail. For the disinterested observer who wishes only the most cursory overview, the final rankings may suffice. But those interested in the game would want much more. Those who were owners of the teams would demand far more information. Those people whose futures depended upon the success of their team would demand a continuous analysis, not only of scores, but also of the individual elements of strategies and programs that lead to success. In the context of school accountability, stakeholders need to know which programs succeeded and which ones failed. In a field littered with "reforms" and "new ideas," some rational method of evaluation other than popularity, enthusiasm or cost must be used. An accountability system that shows policy-makers how intervention strategies correlate with student results can go a long way toward providing such essential program evaluation information. Some of the best practitioners of such systematic evaluation are Robert Slavin and his colleagues at Johns Hopkins University. In his recent book, "Show Me the Evidence!" (1998), Slavin and his colleagues provide examples of how accountability information can be used to monitor program effectiveness. Perhaps the most astonishing aspect of the book is how few programs have been subject to long-term, systematic accountability and analysis. In sports, media commentators are expected to report not only scores, but they must also demonstrate a thorough understanding of the nuances of the game. They routinely debate the meaning of various statistics. The television commentator or sports journalist who simply writes, "Those with winning records are better teams than those with losing records" would soon be out of a job. When it comes to sports, we expect thorough and detailed analysis of all the relevant data. Moreover, we expect the best commentators to look beyond the data and provide insights based on observation, description and qualitative understandings that extend beyond numerical explanations. Though the athletic analogy may seem clumsy, I will celebrate the day when analysis of educational accountability data is taken as seriously by the media and the public as analysis of last weekend's sports games. When that day comes, we all may finally know what works best to help students learn. A comprehensive evaluation of what works best to help students learn must go far beyond mere test scores and include analyses of both how the data reported should be interpreted as well as information about the context in which these results were derived. Beyond the numbers -- Balancing qualitative and quantitative information In educational accountability, numbers are an important part of the story, but they tell only part of the story. The qualitative dimension of accountability -- descriptions, narratives and observations about culture and climate -- creates a lens through which the quantitative data must be viewed. What is the qualitative context of the quantitative data? What are the successes, failures, tragedies and triumphs of this school that help to explain the story behind the numbers? While no principal in America needs one more report to write, a one-page synopsis of a school's qualitative dimension would add greater context to its test scores. Without such a qualitative context, we are left with the sterility of data that, even when presented with abundant statistical complexity, can leave us wondering, "What really happened in that school?" Without the qualitative dimension, our understanding of the "score" is limited, incomplete and possibly inaccurate. What sort of qualitative information should be included? Information about the school climate and environment, the triumphs and tragedies of the school year, and descriptions of any significant changes in programs, personnel or performance can all be expressed in narrative form. Such narrative information is not a substitute for quantitative data, but rather, gives citizens and policy-makers a context in which to interpret numerical results. Meeting the test of common sense An effective accountability system must answer at least four common-sense questions: one about individual student achievement, a second about school performance, a third about ways to help students learn, and a fourth about determining educational effectiveness. In order to provide useful information about student achievement, an accountability system must be based on clear standards that have been communicated to students, parents, teachers and other district stakeholders. Both quantitative and qualitative indicators that measure whether or not these standards have been met must become integral parts of the accountability system. School performance must be based on much more than test scores. Though it is likely to include test data, school performance information must also include how these numbers should be interpreted and the context from which test scores arose. This approach to comprehensive accountability is, to be sure, more challenging than simplistic headlines. It is, however, the only approach that meets the simple test of common sense. References Bracey, G. (October 1999). "The ninth Bracey report on the condition of public education." Phi Delta Kappan, 81 (2), 147-168. Orwell, G. (1949). Nineteen Eighty-four. New York: Harcourt, Brace. Sanders, W.L. (December 1998). "Value-Added Assessment." AASA School Administrator, 55 (11), 24-27. Schmoker, M.J. & Marzano, R.J. (Winter 1999). "Realizing the promise of standards-based education." Educational leadership, 56 (6), 17-21. Schmoker, M.J. (1999a). Results: The Key to Continuous School Improvement. (2nd Ed.) Alexandria, VA: Association for Supervision and Curriculum Development. Slavin, R.E. & Olatokunbo, S.F. (1998). Show me the evidence!: Proven and promising programs for America's schools. Thousand Oaks, CA: Corwin Press. Douglas B. Reeves is the author of "Accountability in Action: A Blueprint for Learning Organizations," (Advanced Learning Press, 2000). He is the president of the International Center for Educational Accountability in Denver. You can continue the conversation with him by calling (800) 844-6599 or via the Internet. |
|||||