"

Chapter 8 – Prepare the Implementation Plan

Learning Goals

Use these learning goals to focus your attention, connect new ideas to your own context, and identify practical ways to apply what you learn. In this chapter you can:

  • Clarify feasible clinical outcomes
  • Design implementation outcomes
  • Build a logic model.

Prepare the Plan

A four-step process diagram: 1. Clarify the Challenge, 2. Assess the Situation, 3. Select the Strategies, 4. Prepare the Plan. Each step is numbered and illustrated.

This is the first of two chapters that bring together behavioural insights, implementation strategies, and an understanding of organisational context into an actionable implementation plan.

 

The focus of this chapter is to choose clinical and implementation outcomes that are feasible and relevant to detect change, and to embed these within a single page logic model.

Clarify Feasible Clinical Outcomes 

To really make a difference, important benefits need to be able to be realised and recognised as improvements.

A critical distinction in implementation is between effectiveness and process outcomes. If a change achieves its clinical benefits, it is seen as effective. Implementation, or process outcomes track and explain why the change did or did not succeed.

  • Effectiveness outcomes confirm that the intervention/innovation made a difference
  • Implementation outcomes describe how the plan was followed.

Understanding this difference is important, when it comes to explaining what happened.

  • A project may have strong evidence but fail to deliver due to poor implementation. Telerehabilitation is an effective alternative to in-person rehabilitation. To experience these benefits, robust and supported technology is required. When technology fails, the benefits of telerehabilitation may not be realised.
  • A project may be well implemented but not achieve any benefits because the evidence was never going to be effective. A tomato diet could be clearly explained and supported so that individuals understand and adhere to it, but it is unlikely to demonstrate benefit as the evidence is not there!

Both types of outcomes are essential. The need to have underlying evidence was introduced early, so that there is a real clinical benefit to be achieved. Process outcomes provide early indicators of progress, while effectiveness outcomes are usually measured later to confirm that the expected clinical benefits have been achieved.  

Effectiveness outcomes

Defining the expected benefits is usually more straightforward than achieving them. Implementation is reported as effective when an evidence-informed intervention or an innovation is introduced into practice and the expected clinical benefits are realised. Effectiveness outcomes refer to these tangible benefits that often refer to changes in patient outcomes.

 

Typically, effectiveness is measured in relation to observable benefits in the way individuals (patients and staff) or aspects of the organisation work. These measures are usually identified during evidence selection and refined through stakeholder input. Ideally, effectiveness is measured within the local context.

Implementation outcomes

The process of change to realise these benefits is usually more difficult to describe and measure. Implementation outcomes indicate how well a new practice, intervention, or innovation is being adopted and embedded into a specific setting. They connect the implementation strategies with effectiveness outcomes, in order to understand the success (or not) of the implementation process.

A simple set of comparisons is offered in the following table.

Aspect 
Effectiveness Outcomes 
Implementation Outcomes  

Definition 

The intended benefits of introducing research or innovation into practice.

The results of changing practice to introduce research or innovation.

Focus 

Does it work?

Can it be adopted, used, and sustained?

Purpose 

Assess whether the change achieves its intended benefit.

Assess whether the change process was successful.

Typical Measures 

Clinical indicators

Health status

Uptake, adherence, fidelity

Acceptability, feasibility

Clinical, Healthcare Examples 

Reduced infection rates

Improved function

Shorter length of stay

Adoption of new model of care

Acceptability amongst users

Staff fidelity to procedures

Timing of Measurement 

Often longer-term, once implementation is complete

Often measured early and throughout implementation

Why it matters 

Demonstrates the positive impact

Explains why the change worked (or not), and under what conditions

Design implementation outcomes

Individuals need to be supported to change key behaviours within their complex organisational systems. Designing bundles of implementation strategies that address behavioural barriers and organisational readiness, has been the focus of the last two chapters. The next challenge is to ensure that these implementation strategies are working.

 

The Behaviour Change Wheel can assist again. In the previous chapter, key determinants (barriers and enablers) of change have been identified and mapped to the intervention functions. These functions operate through mechanisms of change to produce outcomes.

 

Mechanisms of change describe the underlying working processes of behaviour change. They describe potential explanatory pathways of behaviour change, and this can be strengthened by the consistent use of theory.

 

In practice, mechanisms of change are activated when an implementation strategy is utilised. For example, if “lack of knowledge” is the determinant, “education” enables the mechanism of building knowledge, that leads to behaviour change. Education can be delivered in a variety of ways, using different implementation strategies.

 

The following table demonstrates how determinants describe barriers to change across each of the nine intervention functions, identified in the Behaviour Change Wheel. For each function, mechanisms of change are suggested, together with examples of implementation strategies.

Determinants of Change 
Intervention Function 
Mechanisms of Change 
Implementation Strategies 

Poor knowledge, misunderstanding, lack of awareness

Education 

Increase understanding, correct misconceptions

Workshops, courses to increase knowledge of why specific change is important

Lack of skill, poor technique or low confidence

Training 

Enhance ability by developing skills, know-how & confidence

Practical workshops to learn about & practice specific techniques

Low perceived risk/benefit, ambivalence

Persuasion 

Alter beliefs & motivate through emotional engagement

Information and discussion campaigns to emphasise benefits & sway attitudes

Low motivation, absence of rewards, competing priorities

Incentivisation 

Create expectations by linking positive behaviour to rewards

Provide recognition, praise & rewards for ideal behaviours

Physical, emotional, social barriers

Enablement 

Remove obstacles and enhance capacity to act

Improve access, or provide mentoring & resources

Absence of positive role models

Modelling 

Demonstrate & support social learning

Introduce peer & clinical champions

Lack of environment cues or facilitating resources

Environmental Restructuring 

Alter the physical or social environment to prompt behaviour

Place signs and use virtual prompts & reminders near desired behaviour

Easy access to undesired options

Restriction 

Reduce opportunities for undesirable behaviour

Regulate undesired behaviours

Potential gains from not changing behaviour

Coercion 

Threats of loss/cost for undesired behaviour

Fine, penalties, notifications for not changing behaviour

Therefore, the nine intervention functions of the Behaviour Change Wheel are key to understanding how to measure the progress of implementation strategies. Note how each intervention function targets specific mechanisms of change, that underpin every implementation strategy.

Refine implementation outcomes

Implementation outcomes look beyond whether the intervention “worked” in order to understand how it was received in practice. They describe how well the change process is working by measuring the process of using implementation strategies.

 

Back in 2011, an international working group of implementation researchers identified the eight most important concepts that were used to measure implementation outcomes. (This paper is in the Reference section of this chapter). In practice, most projects only choose to measure three or four of these outcomes. Discussion with your stakeholders will help to identify which are the most important outcomes to understand how your change worked.

 

In the early phase of implementation, it is useful to know whether implementation strategies are acceptable, appropriate and/or feasible. Acceptability considers whether people find the change agreeable. Closely related is appropriateness, which evaluates whether the change is perceived as a good fit for the organisation. Feasibility brings in the practical dimension, questioning whether the new procedure is realistic given current staffing levels, workload, and experience. These three implementation outcomes evaluate how the implementation strategies are perceived by the users.

 

During implementation, it is useful to monitor progress at regular intervals, and to make comparisons between actual progress and the plan. Useful implementation outcomes include adoption, cost and fidelity. Adoption measures the intention to use or the uptake of an implementation strategy. Cost often refers to the time and resources required to introduce the change. Once in place, fidelity examines whether the change is being delivered as intended, and as described in the research evidence. These three implementation outcomes evaluate progress and should be measured at different time periods.

 

Finally, as change has been implemented, the implementation outcomes of penetration and sustainability are often measured. Penetration is often referred to as reach, as it evaluates how well the change has been integrated into everyday practice. It may evaluate the proportion of eligible patients who are receiving the new service. Over time, sustainability becomes critical, as a measure of what has been continued for six months or more. Often, additional implementation supports have been withdrawn, and this is a measure of whether the change is fully embedded in the service.

 

Together, these outcomes provide a structured way to assess not just the success of implementation, but its fit in real-world healthcare settings.

Some commonly utilised implementation outcomes are described in the following table.

Example Implementation Outcomes

Implementation Outcome Purpose Examples
Acceptability Do people find the change agreeable or satisfactory? Do clinicians feel the new protocol fits their values and workflow?
Adoption Is the intervention or innovation being used or taken up? How many staff are using the new template or product?
Appropriateness Is the change perceived as a good fit in this organisation? Do staff report that the change addressed the specific problem?
Feasibility Can the change occur within existing workflows and resources? Is the new procedure realistic given staffing availability & experience?
Fidelity Is the change being delivered as intended, described in the evidence? Are clinicians following all steps of the new process?
Penetration How well is the change integrated into normal practice? What proportion of eligible patients are receiving the new service?
Sustainability Will the change continue over time without extra support? Is the new checklist still in use after 6 months?
Cost What is the financial impact of making this change? What are the resource demands to initiate and maintain this change?

 

For each implementation strategy, consider what is most important and able to be measured to demonstrate that they are working.  

Measure implementation outcomes

Implementation outcomes can be measured using a variety of practical methods, including:

  • Surveys & questionnaires: Use relevant validated tools where possible to harness the benefits from previous researchers, and if there are no suitable tools available, create templates to ask staff and patients how acceptable and feasible they find the change.
  • Interviews & focus groups: Explore experiences and discuss why or how behaviour has changed.
  • Routine data & audits: Track how often the new process is used.
  • Behavioural observations: Watch how the change is carried out, record frequencies, or use relevant checklists.
  • Routine monitoring: Use existing performance dashboards or regular reports.

Further, measurement strategies can be selected to reflect the intended mechanisms. The following table illustrates how to align measurement strategies with their underlying mechanisms of change, using practical examples.

Intervention Function 
Mechanism of Change 
Implementation strategies  
Measurement Strategies  

Education 

Increase knowledge & understanding

Briefings

Workshops

Pre/post knowledge tests

Confidence ratings

Training 

Build physical & psychological skills

Practice sessions Simulation labs

Skill assessments

Competency checklists

Persuasion 

Alter beliefs & attitudes

Motivational interviewing Testimonials

Attitude & belief scales

Goals and intentions

Incentivisation 

Create expectation of reward

Recognition programs Bonuses & awards

Rates of uptake, adoption

Reward & recognition

Enablement 

Remove barriers & increases capacity

Increase staffing

Redesign documents & workflows

Barrier checklists

Log of use of supports

Observe workflow changes

Modelling 

Provide demonstration as proof of possibility

Peer demonstrations Clinical champions Mentoring

Peer influence observation

Diaries of role models, mentors

Environmental Restructuring  

Change the physical or social context

Modify spaces & routines Adjust roster &, prompts

Environmental audits

Observe  adjustments

Restriction 

Limit competing behaviours

Bans

Rules

Frequency of behaviours

Policy adherence

Coercion  

Create expectation of punishment, cost

Penalties

Loss of privileges

Compliance rates

Rates of penalties

 

Ongoing stakeholder engagement helps ensure that selected measurement strategies are feasible and relevant for the local context. Whenever possible, leverage existing data sources to reduce duplication and enhance sustainability.

Plan to be responsive

Often, unexpected and unavoidable changes in the healthcare system influence the delivery of implementation strategies. Therefore, it is important to be able to tap into the desired mechanisms of change and look for alternative implementation strategies to support gradual progression.

 

One of the most tangible tasks in preparing an implementation plan is the creation of a logic model. When clear organisational priorities and unique bundles of implementation strategies are agreed upon by core stakeholders, the critical steps of creating a logic model build consensus and ownership.

Build a Logic Model

Logic models  are widely recognised in implementation as visual tools that map relationships between resources, activities, and intended outcomes. They help to visualise the theory of change, by showing how implementation strategies address barriers, utilise key mechanisms and produce change.

 

Building a logic model requires critical discussion between stakeholders and offers a way to maintain the momentum that was generated as stakeholders agreed on and co-designed implementation strategies, as discussed in the previous chapter. A similar co-design process provides opportunities to test assumptions about what will happen and align actions with outcomes. Practically, the shared purpose for change can be clarified and ways to maintain alignment highlighted.

 

The key components of a logic model guide its creation and integrate the accumulated knowledge and understanding of change. The broad sequence is described, following.

  • Identify the change goal, in terms of the desired improvement, practical benefit and importance of action.
  • Define inputs in relation to required resources, including people, processes and materials.
  • Clarify activities by describing the behaviour changes and the implementation strategies supporting this.
  • Describe outputs, which represent the deliverables of each activity.
  • Map expected outcomes and organise into short-, medium- and long-term benefits.
  • Acknowledge external influences in the broader healthcare ecosystem that may influence how that change will occur

Use this table to help identify each component, in relation to a specific organisation and planned improvement.

Component 
Guiding question 
Examples 

Inputs 

What resources are required?

Staff time, data systems, training materials, technology, materials, partnerships

Activities 

What is changing?

Service delivery, clinical pathways, documentation, communication, education and training

Outputs 

What will be delivered?

Number of sessions held, number of tools used, materials distributed, participants reached & engaged, surveys & audits completed

Short-term Outcomes 

What are the early indicators of this change?

Increased knowledge, confidence, skills

Changes in routines

Compliance to new processes

Medium- term Outcomes  

What indicates change is happening?

Adherence to new processes & policies

Maintenance of knowledge, confidence & skills

Realisation of early benefits

Longer-term Outcomes 

What are the broader benefits?

Realisation of all expected benefits

Addressed the original challenging problem

Logic models are commonly presented as a pipeline, supporting a practical program aim, as in the diagram below.

6 boxes with text. Heading reads "implement digital monitoring for patients with type 2 diabetes". Text in boxes from left to right reads: Inputs- Blood glucose monitors, smart scales, data systems, health coaches. Activities- Distribute equipment, digital app training, recruit health coaches. Outputs- Patients enrolled, data informed decisions, health coaches trained. Short term outcomes- Patient self-monitoring, early intervention for non-adherence. Medium term outcomes- Improved HbA1c levels, timely clinical interventions. Long term outcomes- Reduced hospital admissions, increased quality of life.

Co-design a logic model

Once all components are clear, map the relationships between inputs, activities, outputs and outcomes. Use arrows or organise the flow within the same row on a chart. Demonstrate how implementation strategies will lead to the sequence of short-, medium- and long-term outcomes. This provides a visual theory of change.

 

Ensure that outputs (what is delivered) are not confused with outcomes (what changes). Focus on outcomes that are meaningful, observable and practical. Be mindful that many outcomes are beyond individual control. Share drafts with stakeholders to build a shared understanding.

 

A well-crafted logic model provides a shared reference point. It connects actions with outcomes and explains the theory of how the change is supposed to happen. This makes them ideal for navigating complex healthcare systems and demonstrating the progress of change. They can also form the basis for more detailed evaluation and sustainability plans.

 

Logic models are flexible tools that can evolve as staff and patient needs change. With clear mapping, they make it easier to see where adjustments are required when circumstances shift. This adaptability allows the model to capture both planned strategies and emerging changes. Another sample logic model is provided with an emphasis on change for healthcare staff.  

 

Text in 6 columns. Heading reads "Implement staff wellbeing program to reduce burnout". Text from left to right reads: Inputs. Executive sponsor, wellbeing campions, data systems, capability building. Activities. Baselines assessment, Recruit & train campions, redesign workflows, peer support workshops. Outputs. # completed assessments, # champions trained & involved, # new/altered workflows, # staff attendance & engagement. Short term outcomes. Supportive supervision practices embedded, staff wellbeing space & resources, staff utilisation of supports. Medium outcomes. Reduced sick leave, workflow efficiencies realised, patient & staff satisfaction. peer support networks. Long term outcomes. Reduced burnout, improved staff retention, improved staff wellbeing, feedback & evaluation report.

Chapter nine explores how to track progress, make sense of what unfolds, and adapt to stay responsive in complex systems.

Navigate the Paradox Between Planning and Responsiveness

A thoughtful implementation plan brings structure to improvement. By clarifying outcomes, designing implementation strategies, and mapping activities to results, the conditions for meaningful action are clear.

Arrows pointing to central diagonal line with the word "Planning" above and "responsiveness" below.

 

However, in complex health systems, plans are only ever part of the story.

 

Real progress requires the ability to notice what emerges and to respond accordingly. The strength of a logic model lies not in its certainty, but in its usefulness as a shared framework for learning.

 

T

The best plans are living documents that can guide action without constraining it.

 

Case Scenario: “The Plan Looked Great—Until Reality Hit”

Setting: 

A private health network is set to introduce a new remote monitoring program for patients with chronic heart failure.

Challenge:

Often patients with heart failure experience multiple hospital admissions to manage their changing healthcare needs and medication regimes. Research evidence reports that remote patient monitoring, telemonitoring apps, and hybrid clinical support have shown reductions in hospital admissions and improved quality of life for patients with chronic heart disease.

Initial Action:

A cross-functional planning group co-designed a detailed implementation plan, which:

Defined the desired outcome: Fewer readmissions.

Mapped behavioural barriers: Clinician time pressures and low perceived importance of this patient group. Patients’ limited knowledge and awareness of their own heart function.

Included implementation strategies: Clinical champions and efficiency audits for clinicians. Onboarding education and resources for patients.

Included a simplified logic model: Included inputs (home monitoring programs, new workflows), activities (patient education and onboarding, clinical champions, efficiency audits), outputs (completed sessions, audit reports), and outcomes (reduced readmissions, improved patient adherence to home monitoring, improved community acceptance).

This team built workflow charts, onboarding protocols and training sessions. Yet, within weeks of launch, they encountered delays and breakdowns.

Emerging Tension:

This highlighted the Planning vs Responsiveness paradox.
The team had invested in creating a logical implementation plan, but the complexity of care delivery meant the plan couldn’t anticipate every barrier.

Some patients struggled to use the technology. The central nursing team encountered conflict with some GPs who were not fully aware and supportive of the program.

Turning Point:

The project lead responded to the negative feedback and reframed the plan as a “living document. They facilitated regular check-ins with patients, nursing staff and local GPs, to capture real-time insights. With an understanding of the behavioural change mechanisms, the project team was responsive and introduced progressive improvements. For example, onboarding was simplified for digitally inexperienced patients, and nurses co-designed new communication criteria for the local GPs.

Reflection Questions

  • How do you balance planning with responsiveness when things change?
  • What feedback mechanisms could be introduced to enable more rapid responsiveness?

Key Takeaways

1. Planning needs to account for what success looks like. 

Effectiveness and process outcomes shape the vision and experience of potential benefits.

2. Implementation outcomes can influence the process of change.  

Clearly defined process outcomes that address key mechanisms of change provide a plan to achieve effectiveness outcomes.

3. Logic models build consensus and visualise a shared roadmap. 

Logic models are dynamic and can be updated as internal and external situations change.

Additional Resources & Templates

Understanding Implementation Outcomes 

  • This seminal academic paper describes the key implementation outcomes used in implementation science research. It identifies effectiveness, service and client outcomes in Figure 1.

Understanding Facilitation

  • An extremely comprehensive guide to using facilitation skills to improve healthcare has been developed by the Veteran Affairs Behavioural Quality Enhancement Research Initiative.

Developing a Logic Model

  • The following practical guide from NSW Health offers practical information and strategies.

Available Templates

Access the templates on the author's From Research to Reality webpage on the Mosaic website:

  • Match Behavioural Determinants & Implementation Strategies
  • Choose Strategies to Measure Implementation
  • Build a Logic Model

 

definition

License

Icon for the Creative Commons Attribution-NonCommercial 4.0 International License

From Research to Reality: An Implementation Guide Copyright © 2025 by Sharon Mickan is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License, except where otherwise noted.