Back to Journals » Patient Preference and Adherence » Volume 18

A Scale for Measuring Electronic Patient Engagement Behaviors: Development and Validation

Authors Hou S , Wang X, Zhao Z, Ma Y, Liu J, Zhang Z, Ma J

Received 3 November 2023

Accepted for publication 11 April 2024

Published 25 April 2024 Volume 2024:18 Pages 917—929

DOI https://doi.org/10.2147/PPA.S444633

Checked for plagiarism Yes

Review by Single anonymous peer review

Peer reviewer comments 3

Editor who approved publication: Dr Jongwha Chang



Shengchao Hou,1,2 Xiubo Wang,1 Zizhao Zhao,1 Yongqiang Ma,1 Jing Liu,3 Ziyun Zhang,2 Jingdong Ma1

1School of Medicine and Health Management, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, People’s Republic of China; 2Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, People’s Republic of China; 3Administrative Office, Yuebei People’s Hospital, Shaoguan, People’s Republic of China

Correspondence: Jingdong Ma, School of Medicine and Health Management, Tongji Medical College, Huazhong University of Science and Technology, No. 13 Hangkong Road, Wuhan, 430030, People’s Republic of China, Email [email protected]

Purpose: Advancements in electronic health (eHealth) technology have profoundly impacted patient engagement. This study aimed to develop and validate the Electronic Patient Engagement Behavior (EPEB) scale to measure the conceptual and underlying framework of patient engagement behaviors in an eHealth context.
Patients and Methods: Initial measurement items were generated based on a literature review and qualitative research. Two rounds of surveys, a pilot survey and validation survey, were conducted to evaluate the psychometric properties of the scale.
Results: The EPEB scale consists of 15 items in four dimensions: disease information search, physician-patient interaction, social interaction between patients, and disease self-monitoring. In the pilot survey, the exploratory factor analysis revealed a four-factor model, explaining 69.411% of variance. In the validation survey, the Cronbach’s α coefficient of each sub-scale was 0.865, 0.904, 0.904, and 0.900 respectively. The Spearman-Brown split coefficient of the scale was 0.963. The results of the cross-sex measurement equivalence test indicate that all fit indices met the measurement criteria. The confirmatory factor analysis indicated second-order 4-factor model fit the data well. The EPEB has a good reliability and validity.
Conclusion: The EPEB scale provides a reliable tool for measuring patient engagement behaviors in the eHealth context. The utilization of this scale may yield valuable insights into strategies for enhancing patient engagement and optimizing health outcomes.

Keywords: patient engagement behaviors, electronic health, scale development, evaluation, validation

Introduction

Patient engagement is receiving increasing attention in both scientific literature and everyday healthcare practices.1–3 It mainly refers to active participation of patients in their healthcare journey, demonstrating their active role in activities including making informed decisions and leveraging available resources to effectively manage their health conditions.4,5 The role of patients in healthcare has evolved from a passive presence, where doctors made most decisions, to an active participation in their own care. This shift is part of a broader societal movement towards individual rights and autonomy, highlighting the growing recognition of the importance of involving patients in the decision-making processes regarding their health.6 The healthcare system now extends beyond merely treating diseases to also embracing the unique attributes, values, and experiences of each patient.7–9 Engaging patients as partners in their care is clinically appealing for better health outcomes at lower costs.10–12 Several studies have demonstrated that patient engagement has the potential of improving chronic disease self-management and healthcare quality.13–16 Patients are expected to be more and more involved in the whole healthcare system and numerous efforts have been made to promote and encourage patient engagement in healthcare.17

The rise of electronic health (eHealth) technology has profoundly impacted patient engagement behaviors.18–21 With the widespread adoption of digital health tools such as mobile health apps, telemedicine, and online healthcare communities, patients have greater access to healthcare information and resources to support their health and clinical activities and achieve more convenient and efficient communication with healthcare providers.22–24 It is undeniable that the COVID-19 pandemic has further accelerated the integration of information technology in healthcare delivery.25 As many services increasingly transition online, the provision of medical services is being transformed, enabling patients to rely on eHealth tools to fulfill their healthcare requirements.26

In the context of the ongoing technological transformation within the health care sector, patient engagement has undergone significant evolution, integrating advanced electronic characteristics. This progression is manifested in the behaviors of patients, now identified as electronic patient engagement behaviors (EPEB). EPEB may be defined as a spectrum of proactive actions and practices patients undertake by leveraging electronic modalities and digital resources to facilitate enhanced access to health care benefits. With the growing importance of EPEB, it has become an increasingly significant aspect of the healthcare experience, particularly as patients are expected to use eHealth technologies when seeking healthcare services. This shift underscores the necessity for a comprehensive understanding of EPEB, aiming to establish a well-defined framework and evaluation dimensions. In response, the development of a scale specifically designed to measure patient engagement behaviors in an eHealth context is proposed. This scale would provide healthcare providers and researchers with critical insights into the extent of patient interaction with eHealth technologies and the impacts of these behaviors on health outcomes, thereby guiding strategies to optimize patient engagement in the digital era.

Several existing instruments including the Patient Activation Measure,27 Patient Health Engagement Scale,28 Patient Engagement Index,29 and Patient Engagement in Health Care Questionnaire30 have been developed to measure patient engagement. However, these instruments primarily may not adequately capture the multifaceted dimensions and items reflective of new behaviour traits, such as seeking health information online, attending virtual consultations, and tracking health data using wearable devices. This limitation potentially leads to inaccuracies in measurement and hinders our understanding of the impact and effectiveness of eHealth interventions on patient engagement outcomes. As eHealth technology becomes more deeply integrated into healthcare, it is essential to develop a new scale specifically designed to measure EPEB within this rapidly evolving landscape. Our scale is designed to bridge this gap, offering a more fitting and sensitive tool for measuring how patients engage with their health care in the eHealth context.

With the increasing integration of eHealth technology into healthcare, developing a new scale specifically designed to measure Electronic Patient Engagement Behaviors (EPEB) within this rapidly evolving landscape becomes crucial. Our scale is designed to bridge this gap, offering a more accurate and relevant tool for measuring how people engage with their healthcare in the context of eHealth.

Therefore, there are two research questions in this paper, (1) What are the key dimensions of the concepts of EPEB; (2) What are the psychometric properties of a newly developed scale to measure the EPEB?

Materials and Methods

Scale Development

The EPEB scale was designed as a generic instrument with components aimed at assessing electronic patient engagement behaviors in general outpatient or inpatient settings that leverage electronic health technologies and platforms. Based on the theoretical frameworks of engagement behavior framework1 and engagement capacity framework,31 the researchers procured their items of EPEB scale through a systematic process, which involved the following steps (Figure 1): (a) performing a literature review regarding the concept of patient engagement and electronic health to identify and gather relevant dimensions and items from existing measurement instruments that had been used in previous studies, (b) conducting semi-structured interviews with healthcare professionals and patients for gathering qualitative data and insights regarding behaviors of patient engagement in healthcare, (c) synthesizing the information obtained from the literature review and semi-structured interviews in order to gain a comprehensive understanding of the dimensions and potential items for measuring EPEB, (d) determining the theoretical framework and components of EPEB and initial item pool for measuring EPEB (Figure 2), (e) engaging a panel consisting of a physician with extensive industry experience, a registered nurse, a hospital administrator with knowledge in hospital management, a healthcare researcher, and a patient to revise and assess the clarity and appropriateness of the measurement items. The physician and nurse provide clinical insights, the administrator offers operational viewpoints on integrating eHealth, the researcher ensures academic rigor and relevance, and the patient represents the end-user experience. (f) conducting a pilot testing phase to evaluate the effectiveness of the developed EPEB scale in practical settings. A convenience sample of patients was enrolled in the pilot survey to verify the availability of the initial scale and further reduce the item pool.

Figure 1 The process of developing the measured items for the EPEB scale.

Figure 2 The EPEB scale components.

The initial EPEB scale in this study consisted of 20 items divided into four subscales: (1) disease information search (four items), (2) doctor–patient interaction (nine items), (3) social interaction among patients (four items), and (4) disease self-monitoring (three items). A 5-point Likert rating scale was adopted to measure the responses due to its simplicity for respondents’ understanding, and suitability for statistical analysis, and alignment with common survey practices. Ranging from “strongly disagree” to “strongly agree”, this rating scale enabled participants to express their degree of agreement or disagreement with each item. The total score on the scale was calculated by summing the scores of all the items. Higher total scores indicate a higher level of electronic engagement in healthcare. This summation provides an overall measure of participants’ level of engagement, enabling researchers to compare and analyze responses across different individuals and groups.

Validation Framework

The validation framework consists of two studies to ensures a comprehensive validation process. In Study 1, a pilot survey was conducted to determine the dimensions of the EPEB using statistical methods and to further reduce the initial measuring items. The questionnaire is evaluated for ensuring understanding of questions, answer options, and length of items are evaluated. The methodologies employed included reliability testing to assess the consistency of the scale, exploratory factor analysis to uncover the underlying structure of the data, and Rasch analysis to evaluate the item responses for their alignment with the latent trait being measured. Subsequently, in Study 2 the final scale was administered to a larger sample of participants to assess its psychometric properties. The psychometric evaluation included item analysis, reliability tests, validity tests, and cross-sex measurement equivalence tests.

Participants and Data Collection

The data for the pilot survey were collected between April and May 2022 at a leading hospital in Wuhan, China. For the validation survey, data were collected from three first-class hospitals in Hubei Province, China between June and July 2022. Convenience sampling was employed in both studies to ensure efficient data collection. Participants in both studies met the following inclusion criteria: (1) aged 18 years and above, (2) clear-minded with no language communication barriers, and (3) willing to participate in the survey. Following Jackson’s recommendation of a minimum 1:10 sample-to-parameter ratio for maximum likelihood estimation of structural equation modeling,32 the sample size in this study should be no fewer than 200 participants. The pilot survey had a sample size of 312 participants, whereas the subsequent validation survey encompassed a larger sample size of 853 individuals.

Statistical Analysis

Data analysis involved two sample sets. The first sample set was used to conduct reliability tests and exploratory factor analysis (EFA) using IBM SPSS software version 25. To assess the reliability of the initial scale, two measures were employed: corrected item-total correlation (CITC) values and Cronbach’s α coefficient. The following steps were undertaken for the exploratory factor analysis: (1) KMO test and Bartlett’s sphericity test were conducted to determine the suitability of the data for factor analysis. (2) Principal component analysis was performed to extract factors based on the criterion of eigenvalues greater than one. Orthogonal rotation was applied using the maximum variance method.

Additionally, Rasch analysis was conducted as a complementary analysis to provide additional valuable information not fully addressed by EFA. This analysis involved examining item-fit statistics, including infit mean square (InfitMNSQ) and outfit mean square (OutfitMNSQ). By assessing these indices, this study aimed to determine the unidimensionality of the subscales and ensured that the selected items were appropriate in their content relevance. Winsteps software version 3.66.0 was used for Rasch measurement.

The second sample set was analyzed using IBM SPSS software version 25 and IBM SPSS AMOS software version 24. This included item analysis, reliability testing, validity testing, and cross-sex measurement equivalence testing. Item analysis comprised two approaches: the critical ratio method and the correlation coefficient method. These methods allow for a comprehensive examination of the individual items’ performance, highlighting any items that may require further scrutiny or potential removal from the scale. Reliability testing was performed to assess the consistency and stability of the scale. This involved examining both internal consistency, which measures the extent to which the items in each subscale are interrelated, and split-half reliability, which assesses the scale’s reliability by splitting it into two halves and comparing the results. Construct validity, specifically content and structural validity, was evaluated to enhance confidence in the proposed factors derived from the EFA. Content validity was assessed using the S-CVI (scale-level content validity index) and I-CVI (item-level content validity index). Confirmatory factor analysis (CFA) was then performed to test the structural validity and determine whether the data aligned well with the hypothesized factor structure. Finally, to ensure the scale’s applicability across different groups and populations, a multiple-group analysis was conducted using sex as the differentiating factor. This analysis involved both single-group CFA, examining male and female group independently, and multi-group CFA, comparing the factor structure across different sex groups.

Results

Demographic Characteristics

The majority of participants were female (63%), aged between 35 and 50 years (38%), married (74%), and residing in urban areas (83%). Most participants lived with more than three family members (61%) and held a bachelor’s degree (52%). Most participants reported a monthly family income in the range of 3500–5000 yuan (33%). The demographic profile of the participants in the validation survey closely resembled that of the participants in the pilot survey. The majority of participants were female (66%), aged between 35 and 50 years (36%), married (78%), and living in cities (77%). Most participants lived with more than three family members (64%) and held a bachelor’s degree (41%). Most participants had a monthly family income of 3500–5000 yuan (34%).

Pilot Survey

Reliability Test

First, we used a dataset collected from 312 patients (pilot survey). The study conducted a reliability test on the initial scale and purified the initial items according to the following two standards, including deleting items with a CITC not significant (P<0.05) or CITC value lower than 0.50 and removing items whose deleted reliability coefficient (Cronbach’s α) equal or higher than the overall reliability coefficient. The results of the reliability test are shown in Table 1. All items displayed CITC values greater than 0.50. However, after item 6 and item 17 were removed, the Cronbach’s α values of their respective dimensions increased. This item purification process ensured that all dimensions achieved Cronbach’s α values greater than 0.8, indicating a high level of internal consistency for the scale.

Table 1 Reliability Test, Exploratory Factor Analysis and Rasch Analysis of the Pilot Survey

Exploratory Factor Analysis

The results revealed a KMO value of 0.903 and a significant Bartlett’s test value of 4369.407 (P<0.001), confirming the suitability of the data for factor analysis. By conducting principal component analysis, 4 common factors with eigenvalues greater than one were extracted. These factors accounted for a cumulative variance contribution rate of 69.411%, indicating that they successfully explained a substantial portion of the data variability. The 4 extracted factors from the data could be reasonably conceptualized as 4 dimensions: disease information search, doctor–patient interaction, social interaction among patients, and disease self-monitoring.

Each item demonstrated a factor loading ranging from 0.443 to 0.859, indicating strong associations with their respective factors. Moreover, the commonalities of all items exceeded 0.4 (Table 1). However, it was observed that items 5, 6, and 17 displayed cross-loadings greater than 0.4 with differences less than 0.2; thus, these items should be considered for removal from the EPEB scale to ensure the robustness and clarity of the factor structure.

Rasch Analysis

The item fit measure, specifically the mean-square fit statistics (MNSQ infit/outfit), along with the point-measure correlation were examined, and the results are presented in Table 1. In Rasch analysis, items with MNSQ infit/outfit values falling within the range of 0.8–1.4 logits are considered acceptable Among the 20 items initially included in the scale, 17 met this criterion, indicating good fit with the underlying Rasch model. However, it was observed that three items, namely, items 8, 10, and 12, did not meet the acceptable fit criteria. Furthermore, the results of the point-measure correlation (PTMEA-CORR) were satisfactory, as they fell within the desired range of 0.4–0.85. This indicates a reasonable relationship between the items and latent traits being measured. However, further consideration is needed for the three items that do not meet the fit criteria and may require revision or removal from the EPEB scale.

Based on a comprehensive analysis of the reliability test, exploratory factor analysis, and Rasch analysis, the research team decided to eliminate items 5, 6, 10, 12, and 17 from the scale. Subsequently, an exploratory factor analysis was conducted again, and the results are presented in Table 2. The analysis revealed the presence of four distinct common factors, with each item demonstrating a factor loading exceeding 0.6. Furthermore, none of the items displayed cross-factor loadings exceeding 0.4, indicating that they were unambiguously associated with their respective factors.

Table 2 Exploratory Factor Analysis After Removing Unqualified Items

Validation Survey

Item Analysis

In total, 853 patients were included in the validation survey. First, item analysis was used to screen items, including the critical ratio and correlation coefficient methods. (1) Critical ratio method: The obtained samples were sorted by the total score of the EPEB scale from highest to lowest; the top 27% of the scale scores were included in the high group and the bottom 27% of the scale scores were included in the low group. Independent sample t-tests were conducted for the two groups, and the results showed that the mean scores of the items in the high group were higher than those in the low group, and the differences were statistically significant (P<0.001). (2) Correlation coefficient method: The Pearson correlation coefficient method was used to analyze the correlation between each item and the total score. The results showed that the correlation coefficients ranged from 0.632 to 0.802, and the P values were all less than 0.001. There were no items with coefficients less than 0.4 (Table 3).

Table 3 CR and Item-Total Correlation for Each Item of EPEB Scale

Reliability Test

We conducted a reliability test on the scale, including internal consistency and split half reliability: (1) Internal consistency: the Cronbach’s α coefficient of the EPEB scale was 0.932, and the Cronbach’s α coefficient of each dimension was 0.865, 0.904, 0.904, 0.900, respectively (Table 4). (2) Split-half reliability: the results showed that the Spearman-Brown split coefficient of the EPEB scale was 0.963.

Table 4 The Reliability and Convergence Validity of the EPEB Scale

Validity Test

Validity tests, including content validity and structural validity: (1) Content validity: the S-CVI of the EPEB scale was 0.911 and the I-CVI was between 0.833 and 1.000. (2) Structural validity: first- and second-order four-factor structural equation models of the EPEB scale were constructed using AMOS software, and confirmatory factor analysis was conducted. Simultaneously, the four-factor model was compared to single-factor, two-factor, and three-factor models. In the two-factor model, disease information search and doctor–patient interaction were combined as one factor, and social interaction among patients and disease self-monitoring were combined as the second factor. The three-factor model included disease information search as one factor, doctor-patient interaction and social interaction among patients as the second factor, and disease self-monitoring as the third factor.

The specific fitting values of each fitting indices are listed in Table 5, and the results show that the four-factor model had the best fit. Both first- and second-order models met the statistical requirements. A structural diagram of the scale is shown in Figure 3. In addition, the results of the CFA showed that the standardized factor loading of each item was above 0.6, the composite reliability was above 0.8, and the AVE was above 0.5, indicating that the scale had good convergence validity (Table 4). The results of the discriminant validity analysis showed that the AVE square root of each factor was greater than the correlation coefficient between the factor and other factors and that the internal correlation of the factor was greater than the external correlation, indicating that the scale had good discriminant validity.

Table 5 Fit Indices of Factor Structure of EPEB Scale

Figure 3 Path coefficient of four-factor EPEB scale.

Cross-Sex Measurement Equivalence Test

We conducted a cross-sex measurement equivalence test on the EPEB scale, including single- and multi-group confirmatory factor analyses. First, single-group confirmatory factor analysis was used to test the fitting effect of the four-factor model of the electronic patient engagement behavior scale in male and female patient populations. The results show that the fitting indices of both groups met the standards (Table 6). Second, a multi-group factor analysis was used to test the measurement equivalence between male and female patients. The results show that the fitting indices of the morphological equivalent model (M), weak equivalent model (M1), strong equivalent model (M2), and strict equivalent model (M3) met the measurement standards (Table 6).

Table 6 Cross-Sex Measurement Equivalence Test of EPEB Scale

Discussion

This paper proposed a definition of EPEB and provide a multidimensional conceptual model. To measure EPEB, the EPEB scale finally formed in this study contains 15 self-reported items in 4 dimensions which comprehensively reflected patients’ behavioral characteristics. Two patient samples from Hubei Province, China were used to test these measurement items. Exploratory factor analysis and confirmatory factor analysis were used to determine the final structure and content of the EPEB scale, and its reliability and validity were assessed.

Dimensions of EPEB Scale

The EPEB scale incorporates common components such as information acquisition, physician communication, social support, and self-management, which is consistent with previous research on patient engagement.33–36 The first dimension of the scale, disease information search, reflects the behavior of patients utilizing various internet technologies to search for relevant information regarding diseases and treatments. This fundamental engagement behavior can help patients improve their understanding and awareness of their own diseases, enhance confidence in their ability to care for themselves.37–39 The second dimension, doctor-patient interaction, refers to patients’ communication with their healthcare professionals more easily and efficiently in a combined online and offline manner, including request, consultation, feedback and evaluation.40–43 The third dimension, social interaction among patients, involves patients using the internet or mobile devices to engage in socializing and sharing with others, including exchanging experiences, seeking advice and emotional support in online patient community. This behavior benefit patients’ mental well-being and even compliance with treatment through educational and emotional support from individuals facing similar health issues.44–47 The fourth dimension, disease self-monitoring, highlights behaviors of recording and monitoring patient health status by smart health apps or wearable electronic devices that provide self-care support and improve healthy behaviors. The generated content such as everyday diet, blood pressure, blood sugar, and heart rate, can facilitate to identify abnormal conditions and reduce the risks of disease.48–50 This behavior can help patients feel more empowered in managing their own health.51,52 Overall, these dimensions represent key aspects of patient engagement in the eHealth era.

The EPEB scale also possesses unique features. Previous studies explored the measurement of patients’ active roles in care. For instance, the Patient Activation Measure (PAM) incorporates cognition, skills, actions, and beliefs to assess patient activity. Patient Health Engagement (PHE)53 mainly focus on evaluating psychological effects on patients’ lives. While the Patient Engagement Index (PEI)29 and Patient Engagement in Health Care Questionnaire30 have made progress in revealing the behavioral characteristics of patient engagement, they have certain limitations when considering the influence and changes brought about by the eHealth environment. By contrast, the EPEB scale incorporates the impact of information technology, such as the use of online platforms to communicate with doctors or search for disease-related information, joining online communities for social sharing, and the use of wearing devices or smart apps to monitor personal health conditions.

The Psychometric Properties of the EPEB Scale

The findings demonstrate that the EPEB scale has good reliability and validity, indicating that it is a reliable and valid measurement tool for evaluating patient engagement behavior in an eHealth environment. The CFA results revealed that the four-factor model fit the data better than the single-, two-, and three-factor models, substantiating the multidimensional conceptual model of EPEB proposed in this study. The reliability analysis results showed that the Cronbach’s α coefficients of each subscale were above 0.8, and the total Cronbach’s α coefficient was 0.932, indicating good internal consistency reliability. The Spearman-Brown split-half coefficient was 0.963, indicating excellent split-half reliability. All standardized factor loadings were above 0.6, and composite reliability was above 0.8, indicating good convergent validity. Discriminant validity was established as the AVE for each factor surpassing the correlation coefficients between that factor and the other factors. The results of the measurement invariance test confirmed that the EPEB scale had configural, metric, and scalar invariance across sex, indicating that it is equally applicable to both male and female patients. These findings support the reliability and validity of the scale, and provide robust evidence that it is a valuable tool for measuring patient engagement behaviors in the eHealth context.

Research Implications

The development of the EPEB Scale has significant practical and theoretical implications. The EPEB conceptual model facilitates a comprehensive understanding of the characteristics and strategies of patients engaged in medical-related activities in an eHealth environment. Moreover, this study provides a reliable and effective measurement tool that assists healthcare professionals and managers in identifying and addressing the needs and expectations of individual patients in an eHealth environment. Furthermore, this tool has the potential to help patients understand and improve their status of engagement in health-related activities, thereby enhancing their health status and overall well-being. To achieve these objectives, healthcare providers can implement various interventions and recommendations based on different dimensions of the EPEB model. These include the provision of reliable online health information resources, enhancing online interactions between physicians and patients, establishing effective and friendly peer support networks, and developing personalized and intelligent self-management tools. Ultimately, the EPEB scale may contribute to the delivery of more effective and patient-centered healthcare services in eHealth.

Limitations and Future Work

Although this study developed an effective scale, it had several limitations. First, the concept of EPEB remains relatively underdeveloped and requires further refinement. In addition, rigorous testing for cultural compatibility is necessary to ensure the applicability of the scale across diverse populations. Moreover, the scarcity of relevant literature on EPEB limits the scope of this study and highlights the need for further research to better understand this construct. Second, the selection of hospital patients from Hubei, China, as the primary sample may have restricted the generalizability of the findings to broader patient populations. Furthermore, owing to the study’s exclusive focus on the structural dimensions of the EPEB, the important influencing factors and their interrelationships with other relevant variables were not explored. To address these limitations and advance the field, future studies should consider various approaches. First, a larger and more diverse sample should be employed to enhance the external validity of the scale. Additionally, more comprehensive data collection methods such as in-depth interviews and observational techniques can supplement quantitative measures to capture the motives, attitudes, and barriers to EPEB. Third, examining the predictive value of the scale by exploring its relationships with other patient-related variables such as satisfaction and compliance may strengthen its practical utility. Lastly, cross-cultural or cross-national comparative studies could offer valuable insights into similarities and differences in EPEB across different socio-cultural contexts.

Conclusion

In conclusion, the EPEB scale developed in this study is a reliable and comprehensive tool for measuring patient engagement behaviors in the eHealth context. The development and validation of the scale represents a notable contribution to the growing body of knowledge on patient engagement. It has potential value on developing tailored interventions aimed at promoting patient engagement in the digital era, and it also facilitates a deeper understanding of the influence of eHealth technologies on patient engagement for healthcare providers and researchers. Continued research efforts are crucial to refine and validate the findings, and ultimately translate them into actionable strategies for improving patient-provider relationships, shared decision-making, and overall healthcare quality.

Ethics Statement

This study was approved by the ethics committee of Tongji hospital of Tongji Medical College of Huazhong University of Science and Technology (approval no.TJ-IRB20220667). Informed consent was obtained from the study participants and the guidelines outlined in the Declaration of Helsinki were followed.

Acknowledgments

We appreciated the hospitals involved in this study for their support in data collection.

Funding

The work was supported by the Health Commission of Hubei Province scientific research project (WJ2021Q027).

Disclosure

The authors report no conflicts of interest in this work.

References

1. Gruman J, Rovner MH, French ME, et al. From patient education to patient engagement: implications for the field of patient education. Patient Educ Couns. 2010;78(3):350–356. doi:10.1016/j.pec.2010.02.002

2. Carman KL, Dardess P, Maurer M, et al. Patient and family engagement: a framework for understanding the elements and developing interventions and policies. Health Aff. 2013;32(2):223–231. doi:10.1377/hlthaff.2012.1133

3. Pelletier LR, Stichler JF. Action brief: patient engagement and activation: a health reform imperative and improvement opportunity for nursing. Nurs Outlook. 2013;61(1):51–54. doi:10.1016/j.outlook.2012.11.003

4. Halabi IO, Scholtes B, Voz B, et al. ”Patient participation” and related concepts: a scoping review on their dimensional composition. Patient Educ Couns. 2020;103(1):5–14. doi:10.1016/j.pec.2019.08.001

5. Barello S, Triberti S, Graffigna G, et al. eHealth for patient engagement: a systematic review. Front Psychol. 2016;6:2013. doi:10.3389/fpsyg.2015.02013

6. DeBronkart D. From patient centred to people powered: autonomy on the rise. BMJ. 2015; 350:h148.

7. VanVactor JD. Leveraging the patient‐centered medical home (PCMH) model as a health care logistics support strategy. Leadersh Health Serv. 2013;26(2):95–106. doi:10.1108/17511871311319696

8. Wang J, Yao T, Wang Y. Patient engagement as contributors in online health communities: the mediation of peer involvement and moderation of community status. Behav Sci. 2023;13:2.

9. Hardyman W, Daunt KL, Kitchener M. Value co-creation through patient engagement in health care: a micro-level approach and research agenda. Public Manag Rev. 2015;17(1):90–107. doi:10.1080/14719037.2014.881539

10. Nickel WK, Weinberger SE, Guze PA, et al. Principles for patient and family partnership in care: an American college of physicians position paper. Ann Intern Med. 2018;169(11):796–799. doi:10.7326/M18-0018

11. Richards T, Montori VM, Godlee F, Lapsley P, Paul D. Let the patient revolution begin. BMJ. 2013; 346:f2614.

12. Fleming MD, Shim JK, Yen I, Thompson-Lastad A, Burke NJ. Patient engagement, chronic illness, and the subject of health care reform. Med Anthropol. 2021;40(3):214–227. doi:10.1080/01459740.2020.1820500

13. Graffigna G, Barello S, Bonanomi A. The role of patient health engagement model (PHE-model) in affecting patient activation and medication adherence: a structural equation model. PLoS One. 2017;12:6.

14. Bombard Y, Baker GR, Orlando E, et al. Engaging patients to improve quality of care: a systematic review. Implement Sci. 2018;13(1):98. doi:10.1186/s13012-018-0784-z

15. Marzban S, Najafi M, Agolli A, Ashrafi E. Impact of patient engagement on healthcare quality: a scoping review. J Patient Experience. 2022;9:662634815. doi:10.1177/23743735221125439

16. Lin M, Weng W, Apriliyasari RW, Truong V, Tsai P. Effects of patient activation intervention on chronic diseases: a meta-analysis. J Nurs Res. 2020;28(5):e116. doi:10.1097/jnr.0000000000000387

17. Sieck CJ, Hefner JL, Walker DM, Kurien N, Phelps L, McAlearney AS. The role of health care organizations in patient engagement: mechanisms to support a strong relationship between patients and clinicians. Health Care Manage Rev. 2023;48(1):23–31. doi:10.1097/HMR.0000000000000346

18. Graffigna G, Barello S, Triberti S, Wiederhold BK, Bosio AC, Riva G. Enabling eHealth as a pathway for patient engagement: a toolkit for medical practice. Stud Health Technol Inform. 2014;199:13.

19. Sawesi S, Rashrash M, Phalakornkule K, Carpenter JS, Jones JF. The impact of information technology on patient engagement and health behavior change: a systematic review of the literature. JMIR Med Inform. 2016;4(1):e1. doi:10.2196/medinform.4514

20. Auxier JN, Bender M, Hakojärvi H, Axelin AM. Patient engagement practice within perinatal eHealth: a scoping review. Nurs Open. 2023;10(8):4971–4984. doi:10.1002/nop2.1822

21. Amann J. Democratising healthcare: the role of ehealth technologies in driving patient participation. EMJ Innov. 2017;1(1):40–46. doi:10.33590/emjinnov/10312722

22. Sprau AC, Basil G, Borowksy P, Yoon JW, Wang MY. Patient participation with a mobile phone application for objective activity assessment before and after spinal fusion. Cureus. 2020;12:9.

23. Risling T, Martinez J, Young J, Thorp-Froslie N. Evaluating patient empowerment in association with ehealth technology: scoping review. J Med Internet Res. 2017;19(9):e329. doi:10.2196/jmir.7809

24. Craig SL, Calleja Lorenzo MV. Can information and communication technologies support patient engagement? A review of opportunities and challenges in health social work. Soc Work Health Care. 2014;53(9):845–864. doi:10.1080/00981389.2014.936991

25. Hincapie MA, Gallego JC, Gempeler A, Pineros JA, Nasner D, Escobar MF. Implementation and Usefulness of Telemedicine During the COVID-19 Pandemic: a Scoping Review. J Prim Care Community Health. 2020;11:921841020. doi:10.1177/2150132720980612

26. Vicente MA, Fernandez C, Guilabert M, et al. Patient engagement using telemedicine in primary care during COVID-19 pandemic: a trial study. Int J Environ Res Public Health. 2022;19(22):14682. doi:10.3390/ijerph192214682

27. Hibbard JH, Stockard J, Mahoney ER, Tusler M. Development of the patient activation measure (PAM): conceptualizing and measuring activation in patients and consumers. Health Serv Res. 2004;39(4):1005–1026. doi:10.1111/j.1475-6773.2004.00269.x

28. Graffigna G, Barello S, Bonanomi A, Lozza E. Measuring patient engagement: development and psychometric properties of the Patient Health Engagement (PHE) Scale. Front Psychol. 2015;6:274. doi:10.3389/fpsyg.2015.00274

29. Xu RH, Cheung AW, Wong EL. Development and validation of an instrument to measure patient engagement in Hong Kong special administrative region, China. Patient Prefer Adherence. 2018;12:1667–1675. doi:10.2147/PPA.S171026

30. Wu Q, Ye X, Wu Y, Zhao L. Development and psychometric evaluation of the patient engagement in health care questionnaire. J Nurs Care Qual. 2020;35(3):E35–E40. doi:10.1097/NCQ.0000000000000439

31. Sieck CJ, Walker DM, Retchin S, McAlearney AS. The patient engagement capacity model: what factors determine a patient’s ability to engage? NEJM Catal. 2019;5:2.

32. Jackson DL. Revisiting sample size and number of parameter estimates: some support for the N: q hypothesis. Struct Equation Model. 2003;10(1). doi:10.1207/S15328007SEM1001_6

33. Tzeng HM, Marcus Pierson J. Measuring patient engagement: which healthcare engagement behaviours are important to patients? J Adv Nurs. 2017;73(7):1604–1609. doi:10.1111/jan.13257

34. Kimerling R, Lewis ET, Javier SJ, Zulman DM. Opportunity or Burden? A behavioral framework for patient engagement. Med Care. 2020;58(2):161–168. doi:10.1097/MLR.0000000000001240

35. Sweeney JC, Danaher TS, McColl-Kennedy JR. Customer effort in value cocreation activities improving quality of life and behavioral intentions of health care customers. J Serv Res. 2015;18(3):318–335. doi:10.1177/1094670515572128

36. Phillips NM, Street M, Haesler E. A systematic review of reliable and valid tools for the measurement of patient participation in healthcare. BMJ Qual Saf. 2016;25(2):110–117. doi:10.1136/bmjqs-2015-004357

37. Hearld KR, Hearld LR, Budhwani H, McCaughey D, Celaya LY, Hall AG. The future state of patient engagement? Personal health information use, attitudes towards health, and health behavior. Health Serv Manage Res. 2019;32(4):199–208. doi:10.1177/0951484819845840

38. Lu X. The effects of patient health information seeking in online health communities on patient compliance in china: social perspective. J Med Internet Res. 2023; 25:e38848.

39. Lim HM, Dunn AG, Lim JR, Abdullah A, Ng CJ. Association between online health information-seeking and medication adherence: a systematic review and meta-analysis. Digit Health. 2022;8:579741496.

40. Allen S, Rogers S, Harris R. Socio-economic differences in patient participation behaviours in doctor-patient interactions-A systematic mapping review of the literature. Health Expect. 2019;22(5):1173–1184. doi:10.1111/hex.12956

41. Marino F, Alby F, Zucchermaglio C, Fatigante M. Digital technology in medical visits: a critical review of its impact on doctor-patient communication. Front Psychiatry. 2023;14:1226225. doi:10.3389/fpsyt.2023.1226225

42. Kehl KL, Landrum MB, Arora NK, et al. Association of actual and preferred decision roles with patient-reported quality of care: shared decision making in cancer care. JAMA Oncol. 2015;1(1):50–58. doi:10.1001/jamaoncol.2014.112

43. Timmermans S. The engaged patient: the relevance of patient-physician communication for twenty-first-century health. J Health Soc Behav. 2020;61(3):259–273. doi:10.1177/0022146520943514

44. Panzarasa P, Griffiths CJ, Sastry N, De Simoni A. Social medical capital: how patients and caregivers can benefit from online social interactions. J Med Internet Res. 2020;22(7):e16337. doi:10.2196/16337

45. Kim E, Han JY, Moon TJ, et al. The process and effect of supportive message expression and reception in online breast cancer support groups. Psychooncology. 2012;21(5):531–540. doi:10.1002/pon.1942

46. Judd-Glossy L, Ariefdjohan M, Ketzer J, et al. Considering the value of online support groups for colorectal conditions: perspectives from caregivers and adult patients. Pediatr Surg Int. 2022;38(1):31–42. doi:10.1007/s00383-021-05021-x

47. Amann J, Rubinelli S. Views of community managers on knowledge co-creation in online communities for people with disabilities: qualitative study. J Med Internet Res. 2017;19(10):e320. doi:10.2196/jmir.7406

48. Mizuno A, Changolkar S, Patel MS. Wearable devices to monitor and reduce the risk of cardiovascular disease: evidence and opportunities. Annu Rev Med. 2021;72:459–471. doi:10.1146/annurev-med-050919-031534

49. Cheung CC, Krahn AD, Andrade JG. The emerging role of wearable technologies in detection of arrhythmia. Can J Cardiol. 2018;34(8):1083–1087. doi:10.1016/j.cjca.2018.05.003

50. Rivera-ávila DA, Esquivel-Lu AI, Salazar-Lozano CR, Jones K, Doubova SV. The effects of professional continuous glucose monitoring as an adjuvant educational tool for improving glycemic control in patients with type 2 diabetes. BMC Endocr Disord. 2021;21(1):79. doi:10.1186/s12902-021-00742-5

51. Chiauzzi E, Rodarte C, DasMahapatra P. Patient-centered activity monitoring in the self-management of chronic health conditions. BMC Med. 2015;13:77. doi:10.1186/s12916-015-0319-2

52. Volpato L, Del Río Carral M, Senn N, Santiago Delefosse M. General practitioners’ perceptions of the use of wearable electronic health monitoring devices: qualitative analysis of risks and benefits. JMIR Mhealth Uhealth. 2021;9(8):e23896. doi:10.2196/23896

53. Graffigna G, Barello S. Spotlight on the Patient Health Engagement model (PHE model): a psychosocial theory to understand people’s meaningful engagement in their own health care. Patient Prefer Adherence. 2018;12:1261–1271. doi:10.2147/PPA.S145646

Creative Commons License © 2024 The Author(s). This work is published and licensed by Dove Medical Press Limited. The full terms of this license are available at https://www.dovepress.com/terms.php and incorporate the Creative Commons Attribution - Non Commercial (unported, v3.0) License. By accessing the work you hereby accept the Terms. Non-commercial uses of the work are permitted without any further permission from Dove Medical Press Limited, provided the work is properly attributed. For permission for commercial use of this work, please see paragraphs 4.2 and 5 of our Terms.