BACKGROUND

Using evidence to guide clinical practice is widely recognized as the cornerstone of safe, effective, and patient-centred healthcare.1–3 A recent synthesis of a large body of literature related to implementing evidence in healthcare consistently demonstrated improved outcomes, such as length of stay and mortality, along with a positive return on investment (a metric used to evaluate the efficiency of an investment in healthcare).1 While the terminology of using evidence in clinical practice varies across the literature, and includes evidence-based practice (EBP), evidence-based medicine (EBM), evidence-informed decision-making, or evidence-informed practice (EIP), it is commonly conceptualized as the integration of the best available research evidence, clinical expertise, and patient preferences to optimize outcomes, promote resource stewardship, and uphold professional accountability.4–6

In respiratory therapy, as in many health professions, EIP is embedded within national competency frameworks as a professional expectation.7 Its inclusion signals that all RTs are expected to demonstrate the requisite knowledge, attitudes and behaviours to enact EIP at entry-to-practice. Yet how this competency is interpreted and enacted in practice remains unclear. A growing body of evidence indicates that respiratory therapists (RTs) encounter significant challenges in using evidence in practice. For example, Clark et al.8 assessed students’ and faculty’s perceived self-efficacy in EBP and found low mean scores in participants’ knowledge of EBP (e.g., literature searching, understanding statistical tests, and interpreting syntheses) as well as confidence to use evidence in practice. These findings are not unique and point to a range of barriers.8–17 These include heavy workload demands that restrict the time available for engaging in evidence, experiences of burnout, and systemic issues such as limited access to research resources or organizational support. They also reflect more profession-specific challenges, including feelings of inadequacy linked to how the profession is perceived, the underappreciation of RTs expertise, and limited opportunities for mentorship to build confidence and capacity.8–17 Collectively, these challenges highlight the need for more tailored educational interventions that both build individual skills and account for practice realities. Similarly, research across other health professions indicate that the enactment of EIP is often inconsistent and influenced by individual, contextual, organizational-level factors.18–23 These inconsistencies point to ongoing educational, structural and implementation challenges.

As part of a larger program of research, our team draws on foundational literature, established frameworks from other health professions, and input from knowledge users to conceptualize EIP in respiratory therapy. EIP is a process involving three interrelated components: reflective practice, as the ability to critically evaluate one’s decisions and actions to inform future care24,25; shared decision-making, working with people with lived experience to ensure that care decisions align with their values, needs, and preferences26,27; and research awareness, the skills needed to formulate clinical questions, locate relevant sources of evidence, determine its trustworthiness and usability.28,29 This framework served as a basis for assessing current gaps within the respiratory therapy profession.

Given the resource-intensive nature of developing and implementing educational programming, it is important to first identify which components of EIP warrant targeted interventions in the respiratory therapy profession. A clear understanding of the current knowledge, attitudes, and behaviours related to EIP is necessary to ensure that any future educational strategies (e.g., continuing education offerings, entry-to-practice curriculums) are relevant, feasible, and aligned with the learning needs and practice realities of RTs. Accordingly, the research question guiding this study was, what is the current state of knowledge, attitudes, and behaviours of RTs in Canada related to the core components of EIP?

METHODS

Study Design

We used a cross-sectional survey design to assess the knowledge, attitudes, and behaviours of RTs in Canada related to three core components of EIP: reflective practice, shared decision-making, and research awareness. To minimize response burden, each participant was randomly assigned to complete one of three surveys, each focused on a single EIP component.

To ensure our study addressed issues relevant to practice and professional priorities, we engaged a group of knowledge users as a steering committee throughout the project. Knowledge users are defined as individuals positioned to contribute expertise to knowledge production processes and/or who will influence, administer or be an active user of the research results in education, professional development, practice, or policy.30 In this study, we included the board of directors of the professional respiratory therapy association in Canada (the Canadian Society of Respiratory Therapists [CSRT]) as knowledge users, which included respiratory therapy leaders, educators, clinicians and students (n = 11). They provided advice on the study design, advised on the conceptual framework, and reviewed draft surveys for contextual relevance. This study is reported according to the Consensus-Based Checklist for Reporting of Survey Studies (CROSS).31

Participants and recruitment

We performed a sample size calculation using a 95% confidence level, a 5% margin of error, and a conservative population proportion estimate of 50% (p = 0.5) to ensure generalizability of the findings to the population of approximately 12,291 credentialed RTs in Canada.32 This yielded a minimum sample size requirement of 373 participants per component of EIP (total n = 1119).33 To enhance the likelihood of reaching the target sample size, we sent the invitation to twice the number of participants required based on the sample size calculation per survey (total n = 2238). This strategy was intended to account for anticipated nonresponse and improve the probability of achieving sufficient statistical power.34

We employed a simple-randomized, stratified sampling approach to recruit practicing RTs across Canada. Simple randomization ensured that each member of the population had an equal probability of selection, thereby minimizing selection bias. Stratification was based on province or territory to support proportional geographic representation. Eligible participants included credentialed RTs currently practicing in Canada. We excluded student RTs as the items in the surveys pertained to current practice.

Representatives from the CSRT and the Ordre Professionnel des Inhalothérapeutes du Québec (OPIQ) (the regulatory body for RTs in Québec), who were not affiliated with the research project, managed the randomization and distributed survey invitations to their members via email. In May 2025, the CSRT distributed survey invitations to the randomized RTs in all provinces and territories outside of Québec, followed by three reminder emails sent at two-week intervals. In June 2025, the OPIQ distributed the survey to randomized participants in Québec without any follow-up reminders, in accordance with their organizational protocol. Data collection for all surveys closed in July 2025. This two-organization process ensured linguistic and geographic inclusivity.

Finally, to collect the data, we administered the surveys through SurveyMonkey, which automatically assigned each participant a unique identifier to preserve anonymity. As an incentive, participants who voluntarily provided an email address after completing the full survey would be entered into a randomized draw for one of 30 CAD $100 gift cards or one of five conference registrations. These emails were collected separately from their survey responses. Prizes were allocated according to the stratified sampling strategy.

Instrument

We utilized three distinct surveys, each targeting one of the three components of EIP: reflective practice, shared decision-making and research awareness. The decision to focus on these three components was made following a meeting with the knowledge users, where we presented the rationale for the project and the conceptual framework of EIP. During the meeting, the knowledge users provided feedback and engaged in discussions with the research team, supporting the decision to measure all three components. We then determined that including all components of EIP in a single instrument would be too onerous for respondents. As a result, we opted to modify three thematically aligned but distinct surveys for our purpose and randomize respondents to complete one of the three to minimize response burden.

To investigate the reflective practice component, we adopted the Reflective Practice Questionnaire.35,36 This tool has been used across a range of health professions, including medical students, surgeons, nurses, and allied health professionals and assesses the degree to which individuals engage in critical evaluation of their clinical actions and experiences.35,36 The original questionnaire consists of nine subscales. For the purposes of this study, we selected seven subscales and mapped them to our target domains as follows: “self-appraisal” and “uncertainty” subscales were used to represent knowledge (6 items); “confidence-general,” “confidence-communication,” and “stress” were used to represent attitudes (12 items); and “reflection-in-action” and “reflection-on-action” were used to represent behaviours (8 items). We removed the subscales “job satisfaction” and “reflection with others” as they were not conceptually aligned with our framework. The 26 items are ranked on a 6-point Likert scale.

To investigate the shared decision-making component, we developed a composite survey by integrating items from multiple sources, as no single validated questionnaire captured the knowledge, attitudes, and behaviours comprehensively as determined by our literature review and confirmed in personal communications with researchers working in this area. While this approach was informed by theory and prior measures, it has not undergone full psychometric validation, and therefore, the survey should be considered preliminary. Knowledge was assessed using true/false items adapted from surveys developed by Yen et al.37 and Hoffman et al.,38 both of which were grounded in expert consensus and a literature review (14 items). Attitudes were measured using the 9-item IcanSDM scale by Giguère et al.,39 a recently developed instrument with emerging evidence of validity.40,41 These items were scored using a 0–10 visual analogue scale (8 items). Behaviours were assessed using items adapted from Hoffman et al.,38 based on the patient-oriented Shared Decision-Making Questionnaire (SDM-Q)42 and scored on a 6-point Likert scale (12 items). The resultant survey included 34 items.

To investigate the research awareness component, we adapted the Health Sciences–Evidence-Based Practice (HS-EBP) questionnaire as it was developed with input from diverse healthcare professionals, includes questions related to the practice context (often overlooked in other instruments) and demonstrates strong evidence of reliability and validity across various health professions.43,44 The original HS-EBP consists of five subscales; however, for the purposes of this study we used three of its subscales. The “results from scientific research” subscale was used to represent knowledge (14 items). The “beliefs and attitudes” subscale was used to represent attitudes (12 items), and the “development of professional practice” subscale was used to represent behaviours (10 items). We did not use the remaining two subscales, as they were not directly relevant to the domains under investigation. The 36 items are scored on a 10-point numeric scale.

Furthermore, each of our EIP component surveys also contained the same two open-ended questions: “In your opinion, what are the features of EIP?” and "What topics, training or learning activities would help you improve your ability to use EIP in your work?" Finally, each survey collected the same 12 sociodemographic variables: the respondents’ province or territory they’re currently working, self-identified gender, race,32 language, level of education, age, years worked in respiratory therapy, type of practice setting, geographic setting, public vs. private workplace, and employment status.

The research team and one knowledge user (representing the full steering committee) reviewed and revised the items on each component-specific survey to ensure relevance to the respiratory therapy context and contribute content validity evidence. After reviewing the items, we mounted the draft surveys onto SurveyMonkey and shared them with the broader knowledge user group to conduct a pilot test. We asked the knowledge users to complete the three surveys and provide comments on item clarity, structure, and relevance as part of a pilot phase. Of the 11 knowledge users invited, seven (63%) completed the pilot and submitted comments, which the research team reviewed and discussed.

Using the pilot test data, we conducted preliminary psychometric analyses for each survey. For the research awareness survey, Cronbach’s alpha was calculated for the entire scale and its subscales. Internal consistency was strong overall (α = 0.95), although one item showed a negative item-total correlation (–0.32). We retained this item due to its inclusion in previously validated versions of surveys and the small sample size of the pilot. The reflective practice survey demonstrated high internal consistency (α = 0.96) and values exceeding 0.90 across all subscales. For the shared decision-making survey, internal consistency was excellent, with alpha values of 0.90 for the attitude subscale and 0.97 for the behaviour subscale. For the knowledge subscale, which consisted of dichotomous true/false items, we used the Kuder-Richardson Formula 20 (KR-20) to assess the internal consistency reliability. The coefficient was 0.63, which is considered acceptable for an exploratory study.45 Following the pilot, all survey instruments underwent forward and back translation (English-French) following best practices to ensure linguistic and conceptual equivalence across languages.46 Supplemental File 1 includes the final English and French surveys.

Research ethics

This study was approved by the University of Ottawa’s Research Ethics Board (H-11-24-10931). Informed consent was obtained electronically, with participants required to review the consent document and agree before accessing the survey.

Data analysis

Survey responses were exported from SurveyMonkey and analyzed using IBM Statistical Package for the Social Sciences (SPSS) version 29.0 (SPSS Inc., Chicago, Illinois). We grouped participants based on the EIP component of the survey they received. Only surveys with at least 75% of the survey completed were included in the analysis.

Scoring procedures

We calculated subscale scores for knowledge, attitudes, and behaviours by computing the mean of all items within each domain for each survey. To enable meaningful comparison across surveys, we normalized the scores to a 0–100 scale. We assessed the internal consistency of each survey using Cronbach’s alpha.

Descriptive and comparative analyses

We used descriptive statistics (mean [M] and standard deviations [SD]) to summarize scores for each domain (knowledge, attitudes, behaviours) within the three components of EIP: reflective practice, shared decision-making, and research awareness. We conducted one-way analyses of variance (ANOVA) to assess differences between EIP components for each domain. When assumptions of normality or homogeneity of variance were violated, we used Kruskal-Wallis tests as non-parametric alternatives. We conducted post hoc comparisons using Tukey’s HSD test to examine pairwise differences. We adopted a significance level of α = 0.05 for all statistical tests.

To further contextualize these primary analyses and identify potential equity or practice-specific factors, we conducted exploratory analyses examining whether EIP domain scores varied by sociodemographic and practice-related characteristics. We selected specific variables (gender, race, education level, practice setting, geographic location, age, and years in practice) a priori based on previous literature suggesting their potential influence regarding using evidence as healthcare professionals.47–49 We conducted independent-samples t-tests to compare scores between two-level groups (e.g., gender: self-report woman compared to man/prefer not to answer), and Pearson’s correlations for continuous variables (age, years in practice). Given their exploratory nature, these analyses were intended to generate hypotheses for future empirical research and potentially identify patterns that may inform the design of targeted educational strategies.

Open-ended question analysis

We pooled the open-ended responses from the three surveys into a single dataset and analyzed using qualitative content analysis.50 To address instances where participants listed multiple features, we segmented responses into discrete meaning units and inductively coded these units, grouped into preliminary categories, and iteratively refined into broader themes. Each response from participants was then coded dichotomously for each theme (i.e., 1 = present, 0 = absent), allowing us to calculate frequencies and proportions of respondents who mentioned each feature.

RESULTS

The three surveys were accessed 442 times. After removing non-responses and incomplete data, we analyzed the data of 288 participants (overall response rate = 10.2%). When examined per survey, 84 (11.3% response rate) completed the reflective practice component survey, 107 (14.3% response rate) completed the shared decision-making component survey, and 97 (13.0% response rate) completed the research awareness component survey (Figure 1).

A diagram of a company AI-generated content may be incorrect.
Figure 1.Flow diagram of the study.

Participant Characteristics

The groups showed comparable demographic characteristics, with no statistically significant differences in age, gender, race, or years of clinical experience (Table 1). Participants had a mean age of 42.4 years (SD = 10.8), and the majority self-identified as women (n = 216; 77.4%) and White (n = 214; 77%). Most held either an RT diploma (n =134; 48.0%) or held a diploma and a bachelor’s degree (n = 110; 39.4%) and had an average of 17.6 (SD = 16.0) years of experience. Participants were employed across various healthcare settings, with 44.8% working in tertiary care hospitals. Most participants reported working in urban areas (n = 215; 77.1%) and held full-time positions (n = 208; 74.6%).

Table 1.Participants.
Variable Total (n = 279) Reflective practice (n = 84) Shared decision-making (n = 104) Research awareness (n = 91)
Mean age (SD) 42.4 (10.8) 43.0 (12.2) 42.4 (10.4) 41.8 (9.9)
Gender*
Woman 216 (77.4) 60 (71.4) 80 (76.9) 76 (83.5)
Man 58 (20.8) 22 (26.2) 22 (21.2) 14 (15.4)
Prefer not to answer 5 (1.8) 2 (2.4) 2 (1.9) 1 (1.1)
Race*
White 214 (77) 65 (75.6) 78 (74.3) 71 (76.3)
South Asian 18 (6.4) 5 (5.8) 8 (7.6) 5 (5.4)
East Asian 10 (3.6) 2 (2.3) 4 (3.8) 4 (4.3)
Middle Eastern 11 (4.0) 5 (5.8) 2 (1.9) 4 (4.3)
Southeast Asian 4 (1.4) 1 (1.2) 1 (1.0) 2 (2.2)
Indigenous§ 6 (2.2) 2 (2.3) 2 (1.9) 2 (2.2)
Latin American 4 (1.4) 1 (1.2) 2 (1.9) 1 (1.1)
Black 2 (0.7) 2 (2.3)
Prefer not to answer 13 (4.7) 3 (3.5) 6 (5.7) 4 (4.3)
I do not know 1 (0.4) 1 (1.0)
Another race 1 (1.0)
Education level (%)
RT Diploma 134 (48.0) 42 (50.0) 50 (48.1) 42 (46.2)
Bachelor and Diploma 110 (39.4) 37 (44.0) 43 (41.3) 30 (33.0)
Graduate (Master + Doctorate) 28 (10.0) 3 (3.6) 10 (9.6) 15 (16.4)
Other (e.g., Certificates) 7 (2.5) 2 (2.4) 1 (1.0) 4 (4.4)
Years in practice (mean, SD) 17.6 (16.0) 18.7 (12.6) 17.9 (10.9) 16.2 (10.0)
Primary practice setting
Tertiary care hospital 125 (44.8) 32 (38.1) 55 (52.9) 38 (41.8)
Community hospital 74 (36.5) 22 (26.2) 26 (25.0) 26 (28.6)
Community/primary care 40 (14.3) 15 (17.9) 12 (11.5) 13 (14.3)
Higher education institution 12 (4.3) 4 (4.8) 4 (3.8) 4 (4.4)
Outpatient clinic 15 (5.4) 6 (7.1) 3 (2.9) 6 (6.6)
Medical device/pharmaceutical 5 (1.8) 2 (2.4) 1 (1.0) 2 (2.2)
Other 8 (2.9) 3 (3.6) 3 (2.9) 2 (2.2)
Province
Ontario 104 (37.3) 28 (33.3) 36 (34.6) 40 (44.0)
Québec 56 (20.1) 13 (15.5) 23 (22.1) 20 (22.0)
Alberta 37 (13.3) 15 (17.9) 11 (10.6) 11 (12.9)
British-Columbia 28 (10.0) 7 (8.3) 12 (11.5) 9 (9.9)
New Brunswick 17 (6.1) 8 (9.5) 6 (5.8) 3 (3.3)
Nova Scotia 15 (5.4) 6 (7.1) 6 (5.6) 3 (3.3)
Saskatchewan 5 (1.8) 1 (1.2) 3 (2.8) 1 (1.1)
Manitoba 9 (3.2) 3 (3.6) 5 (4.8) 1 (1.1)
Newfoundland 3 (1.1) 1 (1.2) 1 (1.0) 1 (1.1)
Prince-Edward Island 3 (1.1) 1 (1.2) 1 (1.0) 1 (1.1)
Territories 2 (0.7) 1 (1.2) 1 (1.1)
Geography
Urban 215 (77.1) 67 (79.8) 80 (76.9) 68 (74.7)
Suburban 48 (17.2) 11 (13.1) 20 (19.2) 17 (18.7)
Rural 13 (4.7) 5 (6.0) 4 (3.7) 4 (4.1)
I do not know 3 (1.1) 1 (1.2) 2 (2.2)
Employment status
Full time 208 (74.6) 58 (69.0) 77 (74.0) 73 (80.2)
Part time 59 (21.1) 24 (28.6) 19 (18.3) 16 (17.6)
Not currently working 12 (4.3) 2 (2.4) 8 (7.5) 2 (2.2)

Primary findings

Mean scores for knowledge, attitudes, and behaviours varied across the three components of EIP: reflective practice, shared decision-making, and research awareness (Table 2). Shared decision-making yielded the highest mean score for knowledge (M = 86.4, SD = 6.4), whereas research awareness had the highest scores for attitudes (M = 81.4, SD = 13.4) and behaviours (M = 78.3, SD = 10.4). In contrast, reflective practice consistently showed the lowest mean scores across all three domains.

Table 2.Mean scores (0-100) by component of EIP and domain.
Component of EIP Knowledge
Mean (SD) [Min-Max]
Attitudes
Mean (SD) [Min-Max]
Behaviours
Mean (SD) [Min-Max]
Reflective practice (n = 84) 40.6 (9.4) [16.7-60.0] 53.3 (9.4) [31.7-73.3] 61.3 (14.7) [30.0-95.0]
Shared decision-making (n = 105) 86.4 (6.4) [64.3-92.9] 39.6 (14.6) [2.5-68.8] 69.5 (16.2) [20.0-100]
Research awareness (n = 91) 61.2 (19.0) [15.0-100] 81.4 (13.4) [35.8-100] 78.3 (10.4) [40.0-100]

One-way ANOVAs revealed statistically significant group differences for knowledge, attitudes, and behaviours (all p < .001) (Table 3). The effect sizes were large for knowledge (η² = .693) and attitudes (η² = .656), and moderate for behaviours (η² = .187). Post hoc analyses using Tukey’s HSD test confirmed significant pairwise differences across all three EIP components for each domain (Supplemental File 2).

Table 3.ANOVA Results and Effect sizes for knowledge, attitudes and behaviours.
Domains F (df) p-value η²
Knowledge F(2, 277) =312.31 <.001 .693
Attitudes F(2, 277) =263.78 <.001 .656
Behaviours F(2, 277) =31.87 <.001 .187

Note: One-way ANOVAs indicated statistically significant group differences across all three constructs. Effect sizes are estimated based on fixed-effect models.

Exploratory analyses

To further contextualize the primary findings, we explored whether knowledge, attitudes, and behaviours varied by sociodemographic and practice-related characteristics (Table 4). We identified several patterns. For reflective practice, women reported higher behaviour scores compared to men or those who preferred not to answer (M = 63.3 vs. 56.3, p = 0.04). RTs working in suburban settings scored higher than those in urban settings for both knowledge (M = 44.7 vs. 39.5; p = 0.03) and behaviours (M = 70.0 vs. 59.0; p = 0.004). Participants in non-academic settings demonstrated higher scores of attitudes (M = 55.8 vs. 49.9; p = 0.004) and behaviours (M = 64.2 vs. 57.2; p = 0.03) than those working in academic-affiliated institutions.

Table 4.Exploratory findings by EIP component.
EIP Component Domain(s) Variable Differences p-value
Reflective practice Behaviours Gender Women > Men/Another gender
M = 63.3 vs. 56.3
0.04
Reflective practice Knowledge
Behaviours
Geography
(Suburban and Urban)
Suburban > Urban
Knowledge: M = 44.7 vs. 39.5
Behaviours: M = 70.0 vs. 59.0

0.03
0.004
Reflective practice Attitudes
Behaviours
Practice setting
(Non-academic and Academic)
Non-academic > Academic
Attitudes: M = 55.8 vs. 49.9
Behaviours: M = 64.2 vs. 57.2

0.004
0.03
Shared decision-making Behaviours Race
(other than White and White)
Races other than white > White
M = 76.8 vs. 67.5

0.007
Shared decision-making Attitudes Age Older participants reported less positive attitudes
(r = –.26)
< .01
Shared decision-making Attitudes Years in practice Participants with more experience reported less positive attitudes
(r = –.22)
< .05
Research awareness Knowledge Years in practice Participants with more experience associated with higher knowledge
(r = .22)
<.05

For shared decision-making, participants who identified as another race besides White reported higher behaviour scores than White participants (M = 76.8 vs. 67.5; p = 0.007). A weak negative association was also observed between age and attitudes, suggesting that older or more experienced RTs reported less positive attitudes toward shared decision-making (r = –.26; p = < .01 and r = –.22; p = < .05). Finally, we observed no group differences by gender, race, education, or practice setting. However, more years in practice were positively associated with knowledge (r = .22; p = < .05).

For research awareness, we observed no group differences according to gender, race, education level, or practice setting. We noted that a statistically significant positive correlation existed between participants’ years in practice and knowledge scores (r = .22, p < .05). Supplemental File 3 contains all the exploratory statistical results.

Open-ended responses

A total of 268 participants (93.1%) responded to the open-ended question, “What are the features of EIP?” This included 82 participants in the reflective practice group (97.6%), 99 in the shared decision-making group (92.5%), and 87 in the research awareness group (89.7%). We identified nine themes using qualitative content analysis. The most frequent related to research evidence, such as the use of scientific research, guidelines, and hierarchies of evidence as the foundation of practice (228 instances; 79.2%). This was followed by clinical expertise and professional responsibility, referring to clinicians’ knowledge, skills, and accountability in applying evidence (68 instances; 23.6%), and contextual factors, described as acknowledging the influence of healthcare settings, organizational structures, and broader system-level factors have in shaping EIP (49 instances; 17.0%).

Similarly, 251 participants (87.2%) responded to the open-ended question, “What topics, training, or learning activities would help you improve your ability to use EIP in your work?” Response rates by group were 78 in reflective practice (92.9%), 91 in shared decision-making (85.0%), and 81 in research awareness (83.5%). We identified seven themes. The most frequent related to condition- and content-specific application of evidence. Specifically, the desire for training on applying evidence to specific clinical conditions, contexts, or patient populations (85 instances; 29.5%). Participants also described their preferred learning formats, which included journal clubs, webinars, modules, and more (71 instances; 24.7%). Finally, participants had a desire to develop skills to better engage with research, such as understanding statistical tests and types of different methodologies. Specifically, training to build knowledge and technical skills in locating, interpreting, and appraising research evidence (61 instances; 21.2%). (Table 5).

Table 5.Qualitative content analysis.
Theme Description n (%) Example quotes
Open-ended question: What are the features of EIP?
  1. Research evidence
Use of scientific research, guidelines, and hierarchies of evidence as the foundation of practice. 228 (79.2)
  • Involves ongoing learning and research into daily work.
  • EIP to me is having relevant research evidence that helps dictate someone’s practice. The evidence-based research allows consistency in the care provided and provides a standard of care that is proven to work while still allowing us to ask questions and alter when needed.
  • A practice that had gone through rigorous study to prove it's efficacy.
  1. Clinical expertise and professional responsibility
Clinician knowledge, skills, and sense of accountability in applying evidence. 68 (23.6)
  • EIP collates research, practitioner experience, patient feedback and other specific instance related issues to come up with policy.
  • Using the latest research, my own experiences, and the situation to come up with the best course of action.
  • A combination of evidence and personal experience in our practices.
  1. Contextual factors
Acknowledging the influence of healthcare settings, organizational structures, and broader contextual factors that shape the enactment of EIP. 49 (17.0)
  • Forming policy and procedures for practices based on the newest/most evidence backed research for best patient care and outcomes.
  • Protocols with room for discretion, patient safety outcome evaluation
  1. Integration and adaptation of evidence
Combining research, expertise, and other forms of knowledge, and adapting them to patient or situational needs. 30 (10.4)
  • Use of the best available evidence, integration of professional expertise, considerations of client values and preferences, contextual relevance, continuous learning , collaboration and interdisciplinary approach, transparency and accountability.
  • EIP, I think, can be shown in many different ways. When studies are done that show better patient outcomes, hospitals and departments can change the way they practice and what their guidelines are
  1. Patient: Paternalistic orientation
Clinician-driven decision-making “on behalf of” patients, reflecting a more traditional paternalistic model of care. 29 (10.1)
  • Using a systematic approach to integrate best available evidence with respiratory practitioner expertise to make the most informed decisions for patient care.
  • Awareness of current evidence, ability to critically evaluate evidence, incorporate evidence into policies/processes, apply therapies to patients as indicated by evidence.
  1. Driven by personal values
The personal beliefs and values of clinicians that influence how they interpret and apply EIP. 25 (8.7)
  • EIP helps the therapist to build high chance of positive outcome towards treatment or management.
  1. Reflective and evaluative practice
Ongoing reflection and critical evaluation of one’s decisions and their outcomes. 25 (8.7)
  • EIP is a practice that, rather than being rigid and unchanging, allows for change […] a practice that will abandon actions or practices that are proven to be ineffective, and is not afraid to question itself in the face of failure or problematic situations.
  • An up to date, evolving, open-minded practice (not stuck in old ways)
  1. Patient: Shared decision-making
Actively engaging patients in collaborative decision-making, respecting their preferences, values, and lived experiences. 23 (8.0)
  • Good collaboration with the patient and the team leads to better results.
  • Incorporating the needs, values, and preferences of the patient is evidence-based practice. Healthcare decisions should be made in a collaborative manner, ensuring that the treatment aligns with the patient’s unique circumstances and wishes.
  1. No response
20 (6.9)
Open-ended question: What topics, training or learning activities would help you improve your ability to use EIP in your work?
  1. Condition- and-content specific application of evidence
Training on applying evidence to specific clinical conditions, contexts, or patient populations. 85 (29.5)
  • Current trends in ventilation, transition from hospital to home, neuromuscular care
  • Ventilation strategies in ARDS, optimal PEEP, extubation in OR & best reversal strategies / protocol for extubating in OR VAP guidelines, post op pain management,
  • Role of race, gender, and socioeconomic status in regard to lung function. Cultural sensitivity. Current therapies for treating lung conditions.
  • Dying with dignity, quality of life, end of life care, MAiD
  • Pulmonary rehabilitation, smoking cessation.
  1. Preferred learning formats
Preferences for training delivery methods, highlighting how professionals want to engage with EIP learning. 71 (24.7)
  • Journal clubs
  • Workshops
  • Simulations
  • Webinars
  • Online modules
  • Case studies
  • Conferences
  1. Skills to engage with research
Training focused on building knowledge and technical skills required to locate, interpret, and appraise research evidence. 61 (21.2)
  • More training on how to interpret research studies and how to design appropriate research questions.
  • How to assess the quality of research and the different types. What is the easiest method to find new respiratory related literature.
  1. Structural and contextual supports
Organizational and structural conditions that enable or constrain professionals’ ability to engage in EIP training and practice. 50 (17.4)
  • A workplace that involves more research, research clubs, or healthcare practitioners proficient in research.
  • Making time for regular staff meetings to discuss EIP.
  • Integrating multiple different health practitioners together for meetings, get togethers, rounds, etc and involve discussions about topics afterwards.
  1. No response
38 (13.2)
  1. Applying and translating evidence
Training directed at applying evidence in practice and sharing it across professional or institutional settings. 32 (11.1)
  • More exposure to clinical studies and how the results may impact practice in the local context (i.e., SES, academic v. rural facilities, resource constraints)
  • Would like to see some review on how to interpret research papers. It is so easy to pull "research" supporting any perspective but knowing how to interpret research and feel confident it is good quality evidence would be beneficial.
  1. Evidence pushed by others
A desire for others (peers, departments, organizations) to package and deliver evidence (e.g., via protocols or summaries), rather than professionals having to self-source. 26 (9.0)
  • More access to current up to date evidenced based research. A data base or hyperlinks to related studies, journals etc. with subscriptions paid by the college.
  • Giving us free access to clinical based apps on our phone so we have easy access when needed.
  • A directory of high quality, evidence based, practice guideline for each area of practice would be beneficial. The directory should be searchable, easily accessible, and continuously updated as evidence evolves.

Note: EIP = Evidence-informed practice; ARDS= acute respiratory distress syndrome; PEEP= Positive end expiratory pressure; OR=Operating room; VAP; Ventilator-associated pneumonia; MAiD= Medical assistance in dying; SES= socioeconomic status

DISCUSSION

The purpose of this study was to describe the current state of knowledge, attitudes, and behaviours of RTs in Canada related to three core components of EIP: reflective practice, shared decision-making and research awareness. Our results show that no component demonstrated consistently high scores across knowledge, attitudes, and behaviours. Even when one domain scored relatively high (for example, knowledge in shared decision-making), the related domains (i.e., attitude and behaviour) of the same component were comparatively low. While reflective practice had the lowest scores across knowledge, attitudes, and behaviours, scores in all three components indicated room for improvement, suggesting that all areas of EIP require strengthening in the respiratory therapy profession. Such variability in scores also suggests that focusing on individual components in isolation may be limiting. Although reflective practice, shared decision-making, and research awareness can be described separately for analytical purposes, in practice they are interdependent and mutually reinforcing. For instance, the ability to critically reflect on clinical decisions is influenced by a healthcare professionals’ awareness of relevant research evidence, and both are enriched when decisions are made collaboratively with patients.24,28,51 Treating the components as entirely separate risks overlooking these synergies and underestimating the complexity of how EIP is enacted in clinical contexts. Therefore, a more integrated approach may be necessary to guide both understanding and the development of educational strategies to equip RTs to enact EIP as a unified process. Participants’ open-ended responses echoed this view, often describing EIP as a relationship between of multiple components, reinforcing that teaching or learning the EIP components in isolation may not fully capture how EIP is enacted.

In the open-ended responses describing the features of EIP, participants placed a disproportionately high emphasis on research evidence (e.g., scientific studies, randomized clinical trials, clinical guidelines, meta-analyses, and hierarchies of evidence), which aligns with the high mean scores for research awareness attitudes in the results, suggesting that RTs in this sample place strong value on research as the central feature of EIP. In contrast, comparatively fewer participants described reflective practice or shared decision-making as central to EIP, which aligns with the lower quantitative scores observed for those components. This finding is not unexpected, as the respiratory therapy profession, like many others, has been influenced by the dominance of hierarchies of evidence and the privileging of population-level data.22,52–54 While this orientation highlights the respiratory therapy profession’s commitment to grounding care in research, it also suggests the importance of broadening what counts as evidence in practice. Supporting RTs to enact EIP may therefore require rebalancing the emphasis on traditional notions of evidence while reprioritizing and teaching the value of other sources of knowledge, including (but not limited to) individual patient experiences and shared decision-making, local and contextual data, Indigenous, experiential, and other ways of knowing.5,6,54–57

Participants’ open-ended responses about training needs further highlighted the importance of tailoring education to context. Most indicated a preference for learning opportunities directly relevant to their own practice environment (e.g., condition- or population-specific), suggesting that generic training in EIP may be insufficient. To develop robust training, educational programming should be designed to facilitate participants to incorporate their own context into learning (e.g., through case discussions, assessments, or presentations).58–61 Participants also identified a wide range of preferred learning formats (e.g., journal clubs, webinars, modules), which aligns with broader EBP literature on the importance of multifaceted, flexible and diverse delivery approaches.58,59 Taken together, these findings suggest that those involved in education (e.g., educators, professional development coordinators) may consider prioritizing practical, context-sensitive opportunities, such as embedding journal clubs into workplace routines, developing modular online content for self-paced learning, or offering case-based workshops on condition-specific applications. Ensuring flexibility in both content and format may help increase uptake and relevance for RTs working across varied settings.

We identified several exploratory trends that warrant future empirical investigation. For example, reflective practice scores appeared to differ by both geography and practice setting, with suburban RTs reporting higher scores than their urban counterparts, and those working in non-academic environments (e.g., rehabilitation centers, community hospitals, primary care) reporting more positive attitudes and stronger behaviours than those in academic settings. These findings raise important questions about how practice context shapes opportunities for reflection. Possible explanations include greater self-reliance in smaller or less resource-intensive environments, differences in caseloads, expectations related to teaching, availability in resources, or team dynamics (e.g., limited space and recognition for reflective practice).62,63 However, given the sample size and response rate, these results should be interpreted cautiously and considered hypothesis-generating rather than definitive. The influence of practice context on reflective practice and other factors among RTs represents an important area for future research.

We also identified variations related to shared decision-making. Notably, participants who identified as another race besides White reported higher behaviour scores than White participants. This trend may reflect the possibility that individuals from minoritized backgrounds place greater emphasis on inclusivity and patient involvement in clinical decision-making, perhaps informed by lived experiences of inequity or marginalization in healthcare contexts.64,65 In contrast, older participants reported less positive attitudes toward shared decision-making, suggesting a generational or cohort effect in these participants, given that shared decision-making represents a more recent paradigm in healthcare practice.51 Older professionals, who may have been trained in a more “biomedical model” care, could be less inclined to view shared decision-making as central to their professional role.66–68 Likewise, years in practice were associated with less positive attitudes toward shared decision-making, paralleling the findings related to age. This trend could reflect professional norms within healthcare over time, in which more experienced clinicians perceive shared decision-making as impractical or time-consuming in certain clinical settings, despite evidence suggesting otherwise.69 Together, these findings highlight the potential influence of demographic and experiential factors on RTs’ engagement with shared decision-making. More targeted research regarding how shared decision-making is defined and interpreted amongst RTs would be beneficial.

Overall, the findings reveal a consistent pattern in how RTs engage with EIP. Although RTs value research evidence and view it as central to their professional role, this emphasis often overshadows the reflective and shared decision-making dimensions of EIP. The lower scores in reflective practice and shared decision-making, along with the open-ended comments that privileged research evidence, suggest that EIP is often interpreted as applying external knowledge rather than as an integrated process that combines reflection, context, and patient partnership. Strengthening EIP in practice may therefore require rebalancing these dimensions so that reflective practice, shared decision-making and research awareness are treated as interdependent rather than distinct activities.

Strengths and limitations

This study has several strengths. We included a rigorous stratified random sampling strategy to minimize selection bias and obtain a more generalizable sample that better reflects the distribution of RTs across Canada. To measure the three core components of EIP, we adapted three thematically linked surveys with some evidence of validity that were endorsed in collaboration with knowledge users. Another strength was involving with knowledge users throughout the survey design process (e.g. conceptualization, instrument adaptation). Their feedback enhanced both the relevance and content validity of the surveys. Finally, a high proportion of participants who responded to the survey also provided open-ended responses, allowing for both quantitative and qualitative insights to the research question.

No study is without limitations. The surveys used in this study warrant consideration. For reflective practice and research awareness, we adapted the two measures from existing instruments, which had undergone psychometric testing in other health professional populations, which may limit their direct applicability to the respiratory therapy population. Similarly, the reflective practice survey was adapted by selecting and re-mapping subscales to align with our framework. Although this process enhanced conceptual fit, it may have affected validity and reduced comparability with other studies that used the original subscales. Furthermore, given the limited availability of comprehensive measures for shared decision-making, we developed a composite survey informed by theory and existing items. While we constructed each of these surveys according to best practices to ensure their content validity and internal consistency analyses provided preliminary evidence of reliability, each requires more extensive psychometric validation before broader use.

Another limitation is the low overall response rate, which may limit the generalizability of the findings. However, we compared our sample with national registry data and found that participants were broadly similar to the Canadian respiratory therapy population in terms of age and self-reported gender. For example, the proportion of women in our study (77.4%) closely matched national statistics.32,70–72 A separate study of Canadian RTs using convenience sample also showed similar patterns for race and employment characteristics.16 These comparisons provide some reassurance that our sample resembles the Canadian RT population with respect to age, gender, race and employment characteristics, but caution remains warranted when interpreting the findings. Furthermore, how the surveys were administered were not uniform across Canada. Participants in Québec received only one invitation to complete the survey due to local policy, whereas multiple reminders were used elsewhere. Finally, although we report several exploratory findings related to EIP among RTs, these should be interpreted with caution given the response rate and the exploratory nature of the analyses. Future focused empirical studies are needed to confirm the extent of these findings.

CONCLUSION

Using a simple-randomized, stratified sampling approach, we assessed the current state of the knowledge, attitudes, and behaviours of RTs in Canada related to three core components of EIP: reflective practice, shared decision-making, and research awareness. No component demonstrated consistently high scores, and reflective practice was consistently lowest across all domains. Participants emphasized research evidence as central to EIP, while patient and contextual factors were less frequently acknowledged, highlighting the need to broaden how evidence is understood and applied in practice. Preferences for training emphasized contextual relevance and flexible delivery formats, underscoring the importance of tailoring future educational strategies to RTs’ current practice. Our exploratory analyses indicate that certain sociodemographic and contextual factors may shape how RTs engage with EIP. Together, these findings reinforce the need to strengthen all dimensions of EIP in the respiratory therapy profession and provide a foundation for designing educational interventions and future research that support its enactment in practice.


Funding

MZ is supported by a Banting Postdoctoral Fellowship (#509780) from the Canadian Institutes of Health Research (CIHR), and IDG is a recipient of a CIHR Foundation Grant (FDN#143237).

Competing interests

All authors have completed the ICMJE uniform disclosure form. MZ is the deputy editor of the CJRT, and AJW is an associate editor of the CJRT. None were involved in any decision regarding this manuscript. IDG report no conflicts of interest.

Ethical Approval

This study was approved by the University of Ottawa’s Research Ethics Board (H-11-24-10931).

AI Statement

No generative AI or AI-assisted technology was used to generate this manuscript or its content.

Acknowledgement

We would like to thank Tim Ramsay, PhD for providing statistical advice throughout the project, the Canadian Society of Respiratory Therapists and the Ordre Professionnel des Inhalothérapeutes du Québec for facilitating recruitment, and Sébastien Tessier, RRT, MBA for providing feedback on the draft survey items. We would like to thank all participants for taking the time to respond to the survey.