24 Mary P. Bourke, PhD, RN, MSN and Barbara A. Ihrke, PhD, RN Evaluation is a means of appraising data or placing a value on data gathered through one or more measurements. Evaluation also involves rendering judgment by pointing out strengths and weaknesses of different features or dimensions. Evaluation is the “implementation form for accountability as well as one of the basic ways of assuring quality” (Kai, 2009, p. 44). Rossi, Lipsey, and Freeman (2004) describe evaluation as judging a performance based on selected outcomes and standards. In education, evaluation assesses data collected through various methods to measure the outcome of the teaching–learning process. Grading, often confused with evaluation, involves quantifying data and assigning value. Grades serve two purposes. Grades notify students of their achievement as it related to them and others and informs the public of student performance (Shoemaker & DeVos, 1999). Grading represents the achievement of the student as assessed by the faculty or grader. Assessment, in the broadest view, refers to processes that provide information about students, faculty, curricula, programs, and institutions to various stakeholders. More specifically, assessment refers to measures of student abilities and changes in knowledge, skills, and attitudes during and after participation in courses and programs (Angelo & Cross, 1993; Davis, 1993; Gates et al., 2002). Assessment data can be obtained to place students in courses, to provide information about learning needs (see Chapter 2), and to determine achievement in individual courses and programs (see Chapters 16 and 25 to 28). Findings are used to improve student learning and teaching (see Chapter 16) and to improve courses and programs. Educational program evaluation or program review “can be defined as a systematic operation of varying complexity involving data collection, observations and analyses, and culminating in a value judgement” (Mizikaci, 2006, p. 41). Program reviews are typically conducted by the faculty as a self-study (see Chapter 28) and are undertaken to respond to accreditation reviews by state, education, and professional accrediting bodies (see Chapter 29). Accreditation is “a voluntary, peer review process that has been a hallmark of quality in American higher education and professional education for decades” (Grumet, 2002, p. 114). This process serves as a mechanism to ensure the quality of educational programs. Accreditation signifies that an institution, school, or program has defined appropriate outcomes, maintains conditions in which they can be met, and is producing acceptable outcomes (Millard, 1994). According to Alstete (2004), accreditation can be viewed as “a positive, active learning exercise” (p. 3). Accreditation occurs following a period of self-study, evaluation, and periodic review and is primarily focused on the mission of the institution and on student outcomes. Schools of nursing may be accredited by state and regional agencies as well as national nursing organizations. Historically and currently there are two organizations of interest in accrediting nursing education programs: the National League for Nursing Accrediting Commission and the Commission on Collegiate Nursing Education. Effective accreditation programs must be simple, relevant, and cost effective. Regardless of the agency or organization providing accreditation services, nursing faculty must be aware of the standards and participate in the process of their evaluation and review. See Chapter 29 for further information about accreditation of nursing programs. Evaluation is a process that involves the following systematic series of actions: 1. Identifying the purpose of the evaluation 3. Determining when to evaluate 5. Choosing an evaluation design, framework, or model 6. Selecting an evaluation instrument As in the research process, the first step in the evaluation process is to pose various questions that can be answered by evaluation. These questions may be broad and encompassing, as in program evaluation, or focused and specific, as in classroom assessment (Box 24-1). Regardless of the scope of the evaluation, the purpose or reason for conducting an evaluation should be clear to all involved. Formative evaluation (or assessment) refers to evaluation taking place during the program or learning activity (Kapborg & Fischbein, 2002). Formative evaluation is conducted while the event to be evaluated is occurring and focuses on identifying progress toward purposes, objectives, or outcomes to improve the activities, course, curriculum, program, or teaching and student learning. Formative evaluation emphasizes the parts instead of the entirety. The aim of formative evaluation “is to monitor learning progress and to provide corrective prescriptions to improve learning” (Gronlund & Waugh, 2009, p. 8). One advantage of formative evaluation is that the events are recent, thus guarding accuracy and preventing distortion by time. Another major advantage of formative evaluation is that the results can be used to improve student performance, program of instruction, or learning outcome before the program or course has concluded (Gronlund & Waugh, 2009; Sims, 1992). Disadvantages of formative evaluation include making judgments before the activity (classroom or clinical performance, nursing program) is completed and not being able to see the results before judgments are made. Formative evaluation can also be intrusive or interrupt the flow of outcomes. There is also a chance for a false sense of security when formative evaluation is positive and the results are not as positive as predicted earlier. There are many techniques available for formative evaluation of the classroom and program. For example, in the classroom student learning can be measured at a “point in time” using the one-minute paper method. Students are asked to write about the most important points discussed in class and what concepts need further clarification. This technique provides valuable insight into the teaching–learning process. The instructor has an opportunity to clarify information during the next class. For formative evaluation of a program, many schools of nursing use national standardized testing systems such as ATI (Assessment Technologies Institute). Each semester students take a test that identifies the student’s competencies and their placement nationally. This helps to determine student progression through key concepts within the curriculum. Weaknesses within the curriculum can be identified using content-specific testing as the cohorts progress through the nursing program. Thus formative evaluation provides critical data for ongoing changes necessary to improve student outcomes. Summative evaluation (or assessment), on the other hand, refers to data collected at the end of the activity, instruction, course, or program (Grondlund & Waugh, 2009; Kapborg & Fischbein, 2002, Story et al., 2010). The focus is on the whole event and emphasizes what is or was and the extent to which objectives and outcomes were met for the purposes of accountability, resource allocation, assignment of grades (students) or merit pay or promotion (faculty), and certification (Davis, 1994). Summative evaluation therefore is most useful at the end of a learning module or course and for program or course revision. Summative evaluation of learning outcomes in a course usually results in assignment of a final grade. In determining when to evaluate, the evaluator must also consider the frequency of evaluation. Evaluation can be time consuming, but frequent evaluation is necessary in many situations. Frequent evaluations are important when the learning process is complex and unfamiliar and when it is considered helpful to anticipate potential problems if the risk of failure is high. Finally, important decisions require frequent evaluations (Box 24-2). External evaluators are those not directly involved in the events being evaluated. They are often employed as consultants. State, regional, and national accrediting bodies are other examples of external evaluators. The advantage of using external evaluators is that they do not have a bias, are not involved in organizational politics, may be very experienced in a particular type of evaluation, and do not have a stake in the results. Disadvantages of using external evaluators include expense, unfamiliarity with the context, barriers of time, and potential travel constraints. Because evaluators are so critical to the evaluation process, faculty should select evaluators carefully. Box 24-3 lists questions to ask when selecting an evaluator. Evaluation models for nursing education may be found in the nursing literature or may be developed by nurse educators for a specific use. Although evaluation models have been adapted from those used in education (Guba & Lincoln, 1989; Madaus, Scriven, & Stufflebeam, 1988; Scriven, 1972; Stake, 1967) and business, nursing education evaluation models reflect closely the aspects of nursing education and practice that are being evaluated (Billings, Connors, & Skiba, 2001; Germain, Deatrick, Hagopian, & Whitney, 1994; Kapborg & Fischbein, 2002). A model should be selected according to the demands of the evaluation question, the context, and the needs of the stakeholders. Several models are used in nursing evaluation activities; they are described briefly here. For detailed information and an example of the use of one model, see Chapter 28. Chen supports the use of a theory-driven model during evaluation and provides “information on not only the performance or merit of a program but on how and why the program achieves such a result” (Chen, 2004, p. 415). According to Rosas (2005), “a critical requirement of theory-driven evaluation is the development and articulation of a clear theory” (p. 390). Thus the evaluation process flows from a theory-based evaluation of program curriculums or instructional methods. The theory will direct the evaluation process from identifying variables to be measured to the final report (Stufflebeam, 2001). Various theories or models provide the structure for the evaluation processes. Theories can be grounded in the social sciences, in nursing, or in business. McLaughlin and Jordan (as cited in Wholey, Hatry, & Newcomer, 2004) recommend that, before an evaluation is conducted, it is helpful to use a “logic model” approach as an advance organizer. The logic model is a tool that is useful for conceptualizing, planning, and communicating with others about their program. McLaughlin and Jordan further describe one of the first steps in the model as the development of a flowchart that clarifies and summarizes key elements of a program such as resources and other program inputs, program activities, and the intermediate outcomes as well as the end outcomes that the program strives to achieve. It is also capable of showing assumed cause-and-effect linkages among elements in the model. Sanders and Sullins (2006) explain that in order to clarify a program, the logic model describes inputs (resources—fiscal and human—needed to run the program as well as equipment, books, materials), activities (what you are doing), outputs (student demographics, contact hours, assignments, tests), initial outcomes (change in students from activities), intermediate outcomes (longer-term student outcomes), and finally ultimate outcomes (vision for students who have completed the program). This model is extremely helpful when designing a program evaluation. The concepts of the CIPP model (context, input, process, and product) facilitate delineating, obtaining, and providing useful information for judging decision alternatives (Stufflebeam, 1971; Stufflebeam & Webster, 1994). Context evaluation identifies the target population and assesses its needs. For example, the target population of first-year college students is identified and, with surveys, interviews, and focus groups, their needs are assessed. Input evaluation identifies and assesses system capabilities, alternative program strategies, and procedural designs for implementing the strategies. In this case, a college would identify and assess its ability or capacity to start a weekend college program. A plan of action would be designed to implement the new program. Process evaluation detects defects in the design or implementation of the procedure. Product evaluation is a collection of descriptions and analyses of outcomes and correlates them to the objectives, context, input, and process information, resulting in the interpretation of results. Singh (2004) stated that there were “4 key factors in successfully conducting a program evaluation that is based on the CIPP model” (p. 2). The key factors are (1) “create an evaluation matrix,” (2) establish a group to direct the evaluation, (3) “determine who will conduct the evaluation,” and (4) make certain that the evaluators “understand and adhere to the program evaluation standards of utility, feasibility, propriety, and accuracy” (p. 2). The article included several useful tables of questions, data sources, and collection methodologies, thus providing clarity and a systematic design for the use of CIPP for evaluation of nursing programs. Stake’s (1967) countenance model focuses on the goals and observed effects of the program being evaluated in terms of antecedents, transactions, and outcomes (Stufflebeam & Webster, 1994). Antecedents are conditions that exist that may affect the outcome; for example, students with prior college experience will affect the outcome of freshmen scores. Transactions are all educational experiences and interactions, and outcomes are the abilities, achievements, attitudes, and aspirations of students that result from the educational experience. The purpose of this model is to promote an understanding of activities in a given setting. Case studies and responsive evaluations (Stake, 1967) elicit information about the program from those involved in this action-research approach to educational evaluation. In this model, evaluators “interact continuously with, and respond to, the evaluative needs of the various clients, as well as other stakeholders” (Stufflebeam, 2001, p. 63). The Cybernetic Model by Veney and Kaluzny (1991; as cited in Jones & Beck, 1996) is a problem-oriented model focused on immediate feedback within a system. The three phases include: • Phase one: Needs assessment. What are the desired outputs or goals for the program? What is the nature of any problems and what are the expected outcomes for resolution? Determine program strategies and specify goal criteria. • Phase two: Implementation. Is the program progressing as expected? • Phase three: Results assessment. Were the program objectives met? Are program outputs sufficient to justify costs? What is the long-term impact on improvement in health and quality of life? Feedback results are used to make changes to the inputs, outputs, processes, or desired goals. Fourth-generation evaluation is a “sociopolitical process that is simultaneously diagnostic, change-oriented, and educative for all the parties involved” (Lincoln & Guba, 1985, p. 141). Fourth-generation evaluation takes into account the values, concerns, and issues of the stakeholders involved in the evaluation (students, faculty, clients, and administrators). The result is a construction of and consensus about needed improvements and changes (Clendon, 2003). Both the evaluator and the stakeholders are responsible for change. Fourth-generation evaluation incorporates techniques of evaluator observation, interviews, and participant evaluations to elicit views, meanings, and understanding of the stakeholders. As a result, the evaluation becomes not the opinion or judgment of the evaluator but the working toward meaning, understanding, and consensus of all involved in the process. This responsive evaluation informs and empowers the stakeholders for reflection and change. Haleem et al. (2010) determined that a constructivist evaluation approach improved their NCLEX-RN pass rate by 40%. They created a working retreat that involved all faculty in the evaluation process. The goal was to evaluate their entire nursing program and then work on evidence-based solutions to identified problems and thus improve program outcomes. As a group, they evaluated courses, objectives, instruction, curriculum, NCLEX scores, credit loads, courses, sequencing of content, overlap in content, et cetera. The entire process led to an informed, involved faculty. As a result, faculty were stakeholders in the evaluation process and thus collaborators in change. Lessons learned were expressed as “take the time to evaluate the program in a meaningful way, work as a team, and listen to other faculty and the students. These lessons made the difference in promoting student success” (p. 121). Specific changes that resulted in improved scores involved a multifaceted approach. Changes instituted by Haleem et al. (2010) to improve student learning included the following: 1. Ninety percent of the course grade for each clinical course was based on objective testing. 2. Implementing application and analysis-level test items only for nursing exams. 3. Each clinical course was assigned homework to include case studies. 4. All students are required to complete 700 NCLEX-RN practice questions. 5. A workshop was developed that focused on a “good thinking” approach to questions. 6. Comprehensive standardized examinations to be used in each clinical course and at the end of the program. The results are computed as 10% of the student’s grade. 7. Students are required to take an NCLEX-RN review course prior to the NCLEX-RN exam. 8. A new course entitled Boot Camp was developed: a 10-week course meeting 16 hours per week and focusing on content reviews, testing, and remediation. A minimum of 3000 review questions are required for each student. 9. Students are allowed to repeat only one nursing course. If a student receives a C in two nursing or cognate courses, he or she is dismissed from the program.
The evaluation process: an overview
Definition of terms
Evaluation
Grading
Assessment
Program evaluation
Accreditation
The evaluation process
Identifying the purpose of the evaluation
Identifying a time frame
Formative evaluation
Summative evaluation
Determining when to evaluate
Selecting the evaluator
External evaluators
Choosing an evaluation design, framework, or model
Theory-driven model
Using logic models
Decision-oriented models: cipp
Client-centered (or responsive evaluation) models
Naturalistic, constructivist, or fourth-generation evaluation models
Stay updated, free articles. Join our Telegram channel
Full access? Get Clinical Tree