Hes sas22

    Cards (21)

    • Evaluation
      A systematic process that judges the worth or value of something
    • Evaluation provides evidence that what nurses do as educators makes a value-added difference in the care they provide
    • Evaluation ensures staff and students have knowledge, skills, and attitudes that demonstrate competencies to deliver safe, high-quality, evidence-based care
    • Evaluation
      • A process
      • A critical component of other processes (Nursing process, Decision-making process, Education process)
      • A way to provide data to demonstrate effectiveness
      • The bridge at the end of one process that guides direction of the next
    • Steps in Evaluation
      1. Focus
      2. Design
      3. Conduct
      4. Analyze and interpret
      5. Report
      6. Use
    • The results of evaluation are useless if they are not used to guide future action
    • Evidence-based practice (EBP)
      Conscientious use of current best evidence in making patient care decisions
    • Evidence from research
      External; intended to be generalized
    • Internal evidence
      Generated by quality improvement project or EBP with specific population
    • Practice-based evidence (PBE)

      Systematic data collection about client progress generated during treatment to enhance care quality and outcomes
    • Assessment
      Initially gathering, summarizing, interpreting, and using data to decide a direction for action
    • Evaluation
      Gathering, summarizing, interpreting, and using data after an activity has been completed to determine the extent to which an action was successful
    • Evaluation (INPUT) and Assessment (OUTPUT)
    • Determining Evaluation Focus
      • Audience
      • Purpose
      • Questions
      • Scope
      • Resources
    • Evaluation Models
      • Process (Formative)
      • Content
      • Outcome (Summative)
      • Impact
      • Total Program
    • Five Levels of Learner Evaluation
      • Level 0: Learner's participation and readiness to learn
      • Level I: Learner's dissatisfaction and satisfaction during intervention
      • Level II: Learner's performance and attitude in daily setting (Long-term; outcome)
      • Level III: Learner's maintained performance and attitude (Ongoing impact)
      • Level IV: Learner's performance and satisfaction after intervention (Initial process)
    • Designing the Evaluation
      • Needs a level of rigor (precision, exactness, logical organization)
      • Depends on the questions to be answered, the complexity of the scope, and how results will be used
    • Evaluation vs. Research
      Evaluation is audience specific and conducted to make decisions, while research is generic and conducted to generate new knowledge
    • Evaluation Methods
      • What types of data will be collected (complete, qualitative/quantitative)
      • From whom or what will data be collected (participants, caregivers, representatives, documents, preexisting databases)
      • How, when, and where will data be collected (observation, interview, questionnaire, test, record review, secondary analysis)
      • By whom will data be collected (learner, educator, evaluator, trained data collector)
    • Designing the Evaluation: Evaluation Instruments
      • Conduct evaluations using existing instruments when possible
      • Instrument selection steps (literature search, critique, measure performance exactly, documented reliability and validity, affordable, feasible, minimal training)
    • Designing the Evaluation: Evaluation Barriers
      • Lack of focus clarity
      • Lack of ability
      • Fear of punishment or loss of self-esteem
    See similar decks