3 EVALUATION GOALS, OBJECTIVES, AND QUESTIONS The kinds of outcomes considered legitimate in American schools appear to be an ever-changing phenomenon. Every governor's conference, political campaign, professional meeting, and state educational reform brings new objectives and a redistribution of priorities. Unfortunately, all too often there is no redistribution of revenues. Several seemingly contradictory trends are evident. On the one hand, there appears to be a push on the part of society to force the schools back to a basic skill development orientation. The often heated rhetoric about reading, math, and minimum competencies attest to this public concern. From another standpoint the schools appear to be taking over many of the educational responsibilities that were historically considered the prerogative of parents and other socializing agencies of society. Such areas as sex and health education, human relations (including marriage), and the strands of values, morality, and ethics are now addressed in the schools. Overlaying these trends is the desire to develop and enhance higher order thinking skills and problem-solving abilities. The "accountability movement" is rampant in virtually all funding agencies, especially at the state and federal levels. Accountability requires documentation of program impact. The demand for responsive and relevant evaluation is, therefore, ever increasing. With changes in goals and objectives, and accountability, come sociopolitical problems. Educational institutions, from the elementary grades through professional schools, tend to reflect changes in society. Social forces from a variety of political, legal, religious, or economic origins generally find manifestations in curriculum reform, modified instructional systems and professional training programs, or school assessment practices. Many of these forces are at work in today's schools. Such factors as civil rights, the feminist movement, recession/inflation, and consumer awareness impinge on school practice. The bottom line is that in evaluating these changes we had better be asking the right questions or we will get the wrong answers, or we may get answers for questions we didn't ask, or we might find out things we didn't want to know, or . . . well, you get the idea. The importance of involving stakeholders in evaluation question preparation at the earliest possible date, therefore, cannot be overemphasized. 39 40 Designing Program Evaluations IDENTIFYING AND INVOLVING STAKEHOLDERS A great variety of people and organizations will usually be interested in the results of any evaluation. They will vary in the degree of intensity of that interest, but listening to as many of them as possible with help the evaluator frame the right questions. The right questions most likely will lead to the collection of the most relevant information, which in turn will increase the likelihood that these results will be used. Nothing is more frustrating and cost-ineffective than the nonutilization of results. Unfortunately conducting evaluations for the sake of appearance occurs with too great a frequency. These "symbolic" evaluation data are often used as political weapons. Evaluations are most meaningfully done when the results are to be used in explaining ideas, theories, or concepts, or in specific decision-making situations. The variety of potential stakeholders, decision makers, and audiences is well represented in a list compiled by Rossi and Freeman (1989,p. 423). The term stakeholder as used here refers to individuals who have a vested interest in the outcomes of the evaluation. Stated another way, the results of the evaluations will have consequences for the stakeholder. The consequences could be financial, emotional, political, or professional/vocational (Scriven, 1991). • Policy Makers and Decision Makers: Persons responsible for deciding whether a program is to be instituted, continued, discontinued, expanded, or curtailed. • Program Sponsors: Organizations that initiate and fund the program to be evaluated. • Evaluation Sponsors: Organizations that initiate and fund the evaluation. (Sometimes the evaluation sponsors and the program sponsors are identical.) • Target Participants: Persons, households, or other units who participate in the program or receive the intervention services under evaluation. • Program Management: Group responsible for overseeing and coordinating the intervention program. • Program Staff: Personnel responsible for actual delivery of the intervention (e.g., teachers). • Evaluators: Groups or individuals responsible for the design and/or conduct of the evaluation. • Program Competitors: Organizations or groups who compete for available resources. Evaluation Goals, Objectives, and Questions 41 • Contextual Stakeholders: Organizations, groups, individuals, and other units in the immediate environment of a program (e.g., local government officials or influential individuals situated on or near the program site). • Evaluation Community: Other evaluators, either organized or not, who read and evaluate evaluations for their technical quality. A given evaluation may involve only two or three of these groups, but an evaluator can be very surprised by how many groups and individuals may be interested in the results. The evaluator will always be faced with resource-allocation conflict situations. The best professional judgment will need to be applied in deciding on which evaluation question(s) to target. A frequently followed realistic road is that of compromise, as long as professional integrity is not subverted. Even with a very large budget, employment of every known relative, and all the cooperation in the world, it is impossible to investigate or respond to all potentially relevant evaluation questions. It would seem reasonable to focus on a limited number of questions and do the best possible job on those. KINDS OF EVALUATION QUESTIONS The kinds of questions to be addressed by curriculum and program evaluators will, of course, be dictated by the information requirements of decision makers, and the nature and state of reforms or innovations that are proposed or have been implemented. Some questions might be considered formative, for example, how can we improve the materials used in the elementary mathematics curriculum? Or questions might be summative in nature; for example, should the current approach to the teaching of writing be continued? In any event, the evaluation question should be based on objectives or goals. We need goals and objectives to help us frame the right evaluation questions. Following is a list of sample evaluation questions that might be asked. The list and kinds of questions are limited only by the creativity of the evaluator (Payne, 1982a). Focus Category Sample Evaluation Question 1. General Needs Assessment Are the general system objectives in mathematics being met in our elementary schools? 42 Designing Program Evaluations 2. Individual Needs Assessment 3. School Services 4. Curriculum Design 5. Classroom Process 6. Materials of Instruction 7. Monitoring of Student Program 8. Teacher Effectiveness 9. Learner Motivation 10. Learning Environment Are the career information needs of our graduating students being met? Are our school psychological services perceived as adequate by students? What effect has the implementation of the new way of organizing the mathematics courses over the school year had on student achievement? Are teachers following the prescribed teaching techniques in using the new Muth Affective Education Program? Is the drug abuse filmstrip/tape program more effective than the current combination of lecture and programmed materials? Is our current performance and records system adequate in identifying those students in need of academic counseling? To what extent has teachers' verbal reinforcement techniques resulted in a decrease in student retention? Has the tracking system based on post-high—school aspirations resulted in changes in learner motivation? What changes in classroom climate, as perceived by students and faculty, have accompanied the introduction of the new position of assistant principal? Evaluation Goals, Objectives, and Questions 11. Staff Development 12. Decision Making 13. Community Involvement 14. Board of Education Policy 15. School Outcomes 16. Resource Allotment 17. Instructional Methods 43 Did last week's staff-development program on creating performance assessments result in improved teacher skills? To what extent have central office decisions in the last 24 months resulted in lower costs and improved student attitude toward school? Is community involvement in the instructional program a good thing? Are system policies effectively communicated to relevant personnel? To what extent are major cognitive, affective, and psychomotor outcomes being accomplished on a school wide basis? Are projected allotments in the budget adequate for anticipated needs? Do students have a positive attitude toward the Goetz Word Processing Program? There is a tremendous variety of potential tasks represented here, and they would require a great variety of measurement techniques: among them traditional multiple-choice (or true-false and essay) achievement measures, questionnaires, surveys, attitude scales, observations, interviews, and perhaps cost-effectiveness analyses. The key to successful evaluation, however, is a systematic process. But we are getting ahead of the story. Before deciding on how to get to where you are going and whether you enjoyed it once you got there, you must decide on where you want to go. Establishing goals and objectives helps in doing that. We are talking about a variety of educational outcomes. School outcomes can be specified at different levels of generality. 44 Designing Program Evaluations LEVELS OF SPECIFICITY IN EDUCATIONAL OUTCOMES Educational outcomes like people, come in all shapes and sizes. There are big ones that at times are so ponderous that they don't say anything and can't move anywhere. There are others so small as to be almost microscopic. Many are so minuscule that you can't see them and are meaningless because they are so small, even nitpicky. "Truth,"is as so often the case, probably falls somewhere in the middle. Outcomes that are useful undoubtedly have enough bulk to make themselves visible and make a statement but not be so small that they become intellectually invisible. One might conceive of outcomes as falling on a continuum of specificity. Figure 3-1 should help visualize the Figure 3-1 Degrees of Specificity of Educational Outcomes individual differences in specificity. It contains a variety of terms that are frequently used to help focus and direct educational efforts. At the very general end we have educational goals like "Become a good citizen." In the middle (Taxonomy of Educational Objectives) (Bloom, 1956) we might have "Applies Archimedes principles of specific gravity in problem solving." At the specific end we have test items: "What domestic animal is most closely related to the wolf?" Note that the spacing of the outcome-related terms is not even, since objectives and categories of objectives are not created equal. Figure 3-1 is not an equal interval scale. It can be seen that goals like those from national educational commissions would be left-the far left-on our continuum, and the ultimate in specificity is the test or performance item on the right. The test task is an actual sample of what we want the student to know or be able to do. If not an actual sample, it is as good an approximation as we can create. C o