Published on in Vol 10 (2023)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/45598, first published .
Involving Health Care Professionals in the Development of Electronic Health Records: Scoping Review

Involving Health Care Professionals in the Development of Electronic Health Records: Scoping Review

Involving Health Care Professionals in the Development of Electronic Health Records: Scoping Review

Review

1Institute of General Practice and Family Medicine (AM RUB), Medical Faculty, Ruhr University Bochum, Bochum, Germany

2School of Nursing, Saint Elisabeth Group GmbH, Catholic Hospitals Rhine-Ruhr, Herne, Germany

3Chair of Didactics and Educational Research in Healthcare, Faculty of Health, Witten/Herdecke University, Witten, Germany

4Faculty of Health Care, Niederrhein University of Applied Sciences, Krefeld, Germany

Corresponding Author:

Theresa Sophie Busse, Dr Rer Med

Institute of General Practice and Family Medicine (AM RUB)

Medical Faculty

Ruhr University Bochum

Universitätsstraße 150

MA 1/158

Bochum, 44801

Germany

Phone: 49 23432 ext 27782

Email: theresa.busse@ruhr-uni-bochum.de


Background: Electronic health records (EHRs) are a promising approach to document and map (complex) health information gathered in health care worldwide. However, possible unintended consequences during use, which can occur owing to low usability or the lack of adaption to existing workflows (eg, high cognitive load), may pose a challenge. To prevent this, the involvement of users in the development of EHRs is crucial and growing. Overall, involvement is designed to be very multifaceted, for example, in terms of the timing, frequency, or even methods used to capture user preferences.

Objective: Setting, users and their needs, and the context and practice of health care must be considered in the design and subsequent implementation of EHRs. Many different approaches to user involvement exist, each requiring a variety of methodological choices. The aim of the study was to provide an overview of the existing forms of user involvement and the circumstances they need and to provide support for the planning of new involvement processes.

Methods: We conducted a scoping review to provide a database for future projects on which design of inclusion is worthwhile and to show the diversity of reporting. Using a very broad search string, we searched the PubMed, CINAHL, and Scopus databases. In addition, we searched Google Scholar. Hits were screened according to scoping review methodology and then examined, focusing on methods and materials, participants, frequency and design of the development, and competencies of the researchers involved.

Results: In total, 70 articles were included in the final analysis. There was a wide range of methods of involvement. Physicians and nurses were the most frequently included groups and, in most cases, were involved only once in the process. The approach of involvement (eg, co-design) was not specified in most of the studies (44/70, 63%). Further qualitative deficiencies in the reporting were evident in the presentation of the competences of members of the research and development teams. Think-aloud sessions, interviews, and prototypes were frequently used.

Conclusions: This review provides insights into the diversity of health care professionals’ involvement in the development of EHRs. It provides an overview of the different approaches in various fields of health care. However, it also shows the necessity of considering quality standards in the development of EHRs together with future users and the need for reporting this in future studies.

JMIR Hum Factors 2023;10:e45598

doi:10.2196/45598

Keywords



Background

The use of electronic health records (EHRs) is increasing worldwide [1,2]. It has been associated with improvements in health care quality and patient safety [3]. In international literature, different terms are used interchangeably to refer to electronic clinical documentation, such as electronic medical records, electronic patient records, or EHRs [4]. In this paper, we use the term EHRs to refer to different types of electronic documentation of patient health data. EHRs are digitized medical records used in clinical health care within an organization [5]. EHRs are linked to organizations (eg, an EHR that is used by the staff in an intensive care unit [ICU] of a hospital), as opposed to personal health records. Personal health records are characterized by the fact that patients can manage them themselves and provide access to others [6]. EHRs can electronically gather and record both administrative and health-related information as well as store, transmit, and display information from various sources [7]. Traditionally, health-related information in EHRs includes a medical history and medication orders, vital signs, or laboratory results [3,8]. Administrative information may include age, sex, or International Classification of Diseases codes [9]. Depending on the context, EHRs include different submodules, such as medication display, the display of vital signs, or diagnostic information [10]. For example, different content is more critical for work in an ICU than for work in a palliative care unit. Depending on the context, there are EHRs specific to each area of medical care to ensure optimal documentational support [11]. It is useful to ensure that information can be transferred within the units of a hospital and between health care institutions. However, this diversity of records still poses challenges for interoperability [12].

In recent years, technological progress has led to extreme improvements in the field of EHRs in terms of design and functions [13]. EHRs can be used to minimize costs and workload with the help of shared, location independent, and clear documentation [14] and to improve collaboration and coordination between different professions and individuals [15]. In addition to the digitization of previously paper-based documentation, electronic decision support systems and the use of predefined clinical guidelines and standards can support quality improvement based on the latest health care knowledge [3,16].

However, the solitary implementation and use of EHRs in isolation will not guarantee that the quality of care improves. The literature suggests that EHRs with poor usability or functionality may have unintended consequences for their users and patients [17]. For example, the lack of adaption to workflows [18,19] and user needs [20], poor usability [21], and unstructured data sets in EHRs lead to high cognitive demands on users [22]. These aspects are associated with work-related stress, fatigue, and burnout for the main user groups of EHRs: nurses [23,24] and physicians [25-27]. Furthermore, poor usability has been associated with patient harm [28]. For example, it can make it difficult for health care providers to access necessary medical data for the treatment of patients or lead to misinterpretation of available data. This can lead to misdiagnosis, incorrect treatment, or unsuitable medication for a patient’s condition. This can put the patient’s health and well-being at risk [28]. Therefore, user acceptance is essential for successful implementation, actual use, and user satisfaction of EHRs [29]. Expected usefulness, technical concerns, technical problems, and expected workflow challenges can facilitate or hinder technology acceptance [30].

To promote international joint development projects, globally valid standards have been drawn up. For example, there is International Organization for Standardization (ISO) 9241-210, which focuses on the ergonomics of human-system interaction. It is an important standard for classifying and demanding usability engineering measures in product development processes such as the development of EHRs. Unfortunately, these standards are often only partially complied with, leading to the various abovementioned problems.

In addition, the involvement of users is necessary to adapt EHRs to the needs of health care professionals and to ensure their acceptance [31]. This is increasingly being addressed, resulting in an expansion of projects involving future users in EHR development. Existing reviews have focused on the involvement of users in technology development. In recent years, several reviews have been published to address the involvement of users in the development of different health-related technologies [32,33]. The focus of these reviews has been, on the one hand, on the involvement of different user groups, such as older people [34-39], people living with dementia [40], or patients with chronic diseases [41]. Specific to these groups is the fact that their cognitive and physiological characteristics must be addressed in the development of digital technologies. On the other hand, reviews cover different use cases of technologies such as mobile health [42], serious digital games for health promotion [43], or for the treatment of depression [41]. In addition, other reviews cover more general aspects of user involvement for the development of health-related technologies [32,33]. Despite the empirical evidence supporting the need for user involvement in the development and implementation of EHRs, this topic is largely excluded from the reviews. For example, in 3 reviews covering generic aspects of user involvement, no study focused on EHRs [32,33,44]. Different approaches such as participatory design or co-design are common practices to involve users in the development of new technologies. For example, on the one hand, participatory design actively and creatively involves both users and designers and thus includes different individual qualifications [31,45]. This approach can be defined as “...a process of investigating, understanding, reflecting upon, establishing, developing, and supporting mutual learning between multiple participants in collective ‘reflection-in-action’. The participants typically undertake the two principle roles of users and designers where the designers strive to learn the realities of the users’ situation while the users strive to articulate their desired aims and learn appropriate technological means to obtain them” [45]. On the other hand, an approach such as co-design is defined as an “active collaboration between stakeholders in the design of solutions to a pre-specified problem” [46]. Although these 2 approaches are often used interchangeably and synonymously, they differ in how much choice is given to users in the development of a technology. It can be assumed that the participatory design approach gives users more influence than the co-design approach.

Aim

In existing studies, the design of the methodology varies depending on resources, time period, and technology. When using participatory design or co-design methods, methodological choices must be made [46]. For the planning of similar research projects and a sensible use of diverse methods, it is crucial to provide an overview. Within the framework of a scoping review, we therefore investigated which forms of user involvement have been used to date, under what circumstances, and with what results. The result can also be used to facilitate guidelines for the involvement of health professionals in the development of EHRs. The overview in the metadata table in Multimedia Appendix 1 [47-115] is intended to be particularly helpful in this regard. In addition to the range of possibilities, specific classifications can also be made as to which method is helpful and for which objective.

The review was guided by the question: “How are health care professionals involved in the development of EHRs?”


Overview

The Methods section is reported as recommended by the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 statement [33]. The presentation of this scoping review is based on the methodological specifications by Peters et al [116].

This methodology was developed using an a priori scoping review protocol [117]. The decision to use a scoping review methodology was based on the identification of a gap in current knowledge [117] and the need for an overview of different methods without any assessment. The result should be a narrative account with a focus on the different ways and methods of involving end users in the development of EHRs.

The methodology of scoping reviews is gaining popularity, particularly in the field of health care [118]. Whereas systematic reviews aim to synthesize collate empirical evidence on a focused research question and present the evidence from the reviewed studies [119], scoping reviews map the existing literature on a topic area [120]. In addition, scoping reviews provide a descriptive overview [121] and are therefore an appropriate method for addressing the research question.

This review aimed to provide an overview of the existing ways and methods of user involvement in the development of EHRs in the literature. The four specific objectives of this review were (1) to conduct a systematic search of the published literature for studies focusing on user involvement in the development of EHRs, (2) to present the characteristics and range of methods used in the identified manuscripts, (3) to explore the reported challenges and limitations of the methods, and (4) to make recommendations for the further development of the approach to the development of EHRs and to improve the consistency with which these types of studies are conducted and reported.

Planning for the review began in January 2021. The review was conducted and evaluated from June 2021 to April 2022. Four people were involved in carrying out the review (JL, CJ, SK, and TSB). JL and CJ had experience in conducting (scoping) reviews. CJ, SK, and TSB worked with prospective users to develop an outpatient EHR, an inpatient EHR, and a cross-sectoral EHR for pediatric palliative care with future users.

Eligibility Criteria

Studies were eligible for inclusion if they described the involvement of health care professionals in the development of EHRs. This explicitly included studies that examined a specific EHR. However, excluded studies focused on the general workload resulting from the use of different EHRs in different institutions or other parameters related to different EHRs, regardless of their design. Articles published in languages other than English were excluded. Manuscripts that described a process without performing it were excluded from the scoping review. Gray literature was not included because of the focus on research projects, although this was included in the scoping review methodology [122].

SK and JL formulated the inclusion and exclusion criteria and discussed them with TSB and CJ. The inclusion and exclusion criteria are listed in Textboxes 1 and 2.

Textbox 1. Inclusion criteria.

Languages

  • English

Publication period

  • 2011-2021

Format

  • Full text available

Study design

  • Empirical studies on the development of an electronic health record (EHR) or modules or submodules where health care professionals are involved
  • Needs assessment
  • Requirements testing or evaluation
  • Qualitative, quantitative, and mixed methods studies

Forms of publication

  • Papers published in a scientific journal

Product

  • EHR
  • Submodules or modules integrated into an EHR
  • Same EHR in different stages of development

Development phases

  • Pending testing or a major effectiveness study
  • Implementation
  • Evaluation

Setting

  • All settings in health and social care

Participants in development process

  • Health care professionals, even if other groups of people are involved
Textbox 2. Exclusion criteria.

Languages

  • Languages other than English

Publication period

  • Before 2011

Format

  • Abstract only or full text not available

Study design

  • Reviews
  • Randomized controlled trials

Forms of publication

  • Study protocols
  • Conference papers
  • Gray literature
  • Books
  • Bachelor thesis, master thesis, or similar works

Product

  • Decision support systems
  • Personal health records
  • Other technologies (integrated apps)
  • Comparison of different electronic health records in one survey
  • Electronic health record for education or training purposes
  • Hardware-specific evaluations

Development phases

  • No restriction was made with regard to the development phase

Setting

  • No setting was excluded

Participants in development process

  • Exclusively patients or other users
  • Trainees or students in the health care sector without patient contacts

Information Sources

The search was carried out in the PubMed, CINAHL, and Scopus databases. A supplementary search was carried out in Google Scholar. The final search was performed by JL and SK on March 17, 2021, using the search strings from Multimedia Appendix 2. The forward-and-backward citation tracking [123] was then carried out by JL using Scopus and Google Scholar.

Search Strategy

First, an initial limited search was conducted in a selection of relevant databases to analyze possible terms in the title and abstract to identify keywords describing the articles. This was followed by a search of all databases using all identified keywords. SK and JL formulated the basic idea of the review and conducted the initial searches. Afterward, SK, JL, and TSB developed the search terms. The search terms used were based on two main categories: (1) search terms around the term EHRs and corresponding synonyms as well as Medical Subject Headings terms (PubMed) and subject headings (CINAHL) were used, and (2) search terms around the term participatory design with corresponding synonyms and Medical Subject Headings terms or subject headings were used. The search strings for PubMed, CINAHL, Google Scholar, and Scopus are shown in Multimedia Appendix 2.

The search in Google Scholar used a substantially shortened search string, as the search engine cannot process longer, complex search strings. This resulted in several results that did not meet the inclusion criteria. Therefore, using Google Scholar’s sort by relevance function, only the first 250 results were checked for eligibility, of which 47 were selected.

Selection Process

All citations were imported into the bibliographic manager EndNote (Clarivate), and duplicate citations were automatically removed, with further duplicates removed if found later in the process. The citations were then imported into the software [124] to subsequently check the relevance of the titles and summaries and to characterize the data of the full articles. Rayyan provided blinded checking and automatically displayed matching inclusions, exclusions, and conflicts after blinding was turned off.

First, the titles and abstracts were checked by SK and JL to ensure compliance with the inclusion criteria. Differences were discussed with TSB. Subsequently, TSB and JL screened titles and abstracts for the forward-and-backward citation tracking results, and the differences were discussed with CJ.

All citations deemed relevant after title and abstract screening were obtained for subsequent review of the full-text article. For articles that could not be obtained through institutional holdings that were available to the authors, attempts were made to contact the authors of the source and request the article. In addition, articles were requested via interlibrary loan.

SK and JL screened the full texts; the differences were discussed with TSB. TSB and JL screened the full texts of the forward-and-backward citation tracking results; the differences were discussed with CJ.

At this stage, studies were excluded if they did not meet the eligibility criteria. After reviewing approximately 25 articles independently, the reviewers met to resolve any conflicts and to ensure consistency among the reviewers and with the research question and purpose [125]. The excluded studies were appropriately labeled with the reason for exclusion to improve traceability.

Data Analysis

Categories were formed deductively. This was based on a systematic review by Vandekerckhove et al [33] that focused on electronic health interventions. The aim of the review by Vandekerckhove et al [33] was to report and justify participatory design methods in empirical eHealth studies for further development of the methodology. The decision to follow this review was based on its comprehensive presentation and its fit for the research question pursued here. However, the categories were supplemented by inductive categories that emerged from reviewing the material. The categories can be named as “factual categories” according to Kuckartz [126], designating specific facts in the included studies. All codes were reviewed, coded, and discussed in regular meetings by TSB, CJ, and SK.

Categories for Syntheses

Owing to the diversity of study designs and the research questions, a quality assessment was not performed. Following the approach of Vandekerckhove et al [33], an assessment of the sufficiency and design of reporting was conducted. To improve comprehensibility, the inductive categories were supplemented by key questions (based on the definitions of the categories that were created and constantly refined during the analysis process) and served to represent the collected data items. This was partly based on the categories in the study by Vandekerckhove et al [33], whose review dealt with eHealth interventions. For example, category 1 in this review was developed based on the category “eHealth intervention” by Vandekerckhove et al [33] and category 3—study participants—was based on the category “stakeholder types” by Vandekerckhove et al [33]. Category 4—methods and materials—was based on the category “tools of participatory design” by Vandekerckhove et al [33]. The other categories were derived from the material itself, as described earlier in the Data Analysis section. This resulted in the following division, which was used to structure the method representation:

  1. Focus and scope of the studies: What is mentioned about the characteristics of the EHR to be developed and the stage of the technology (prototype, already implemented EHR)? What was the aim of the studies?
  2. Setting: Where did the involvement in the development of the EHR take place?
  3. Study participants: Who was involved in the development? Which characteristics were mentioned when describing the study participants?
  4. Methods and materials: Which study design was used? Which terminology was used to describe the involvement process? Which methods were used? Are there any physical materials used in the process? How often were participants involved in the process (involvement counts as renewed involvement if it takes place at a later point in time and contributes to the further development of the technology)?
  5. Frameworks, theories, and guidelines: What approaches have been used and influence the choice of methods? Which approaches were used only within the data analysis or a specific method without influencing the choice of methods for the entire study? Which design guidelines were mentioned that influenced the basic logic of the EHR design?
  6. Competencies of the researchers: What competencies do researchers contribute in terms of development?

Study Selection

The study selection is described in Figure 1.

The initial search resulted in a total of 23,446 hits (PubMed: n=8281, 35.32%; CINAHL: n=1846, 7.87%; Scopus: n=13,319, 56.81%). In addition, 47 records were extracted from other sources (Google Scholar) after screening the first 25 pages of approximately 7710 results. From a total of 23,446 hits from the initial search and these 47 additional records, 19,002 (81.04%) hits remained after duplicate reduction.

Of these 19,002 articles, 18,830 (99.09%) articles were excluded during title and abstract screening. The remaining 172 texts were subjected to full-text screening, resulting in a total of 74 titles.

The forward-and-backward citation search yielded a total of 2769 hits (755 by forward citation and 1985 by backward citation). Automatic deduplication reduced the number of hits to 2665. Manual duplicate reduction led to a final result of 2625 hits. After title and abstract screening (34 texts remaining), full-text screening was carried out, resulting in 23 articles.

These 23 articles from the forward-and-backward citation search were included in the final assessment along with the previous 74 studies from the initial search. Of these 97 studies, 27 (28%) studies were excluded because of insufficient information and duplicates. The remaining 70 articles were included in the evaluation and can be found in the metadata table in Multimedia Appendix 1.

Figure 1. Flow diagram of the study selection process. EHR: electronic health record.

Study Characteristics

Multimedia Appendix 1 includes the metadata of the included studies.

Results of Syntheses

Focus and Scope of the Studies

The included studies targeted different objectives in EHR development, which can be divided into 8 groups. Some of the studies focused primarily on eliciting users’ needs and wants toward EHRs. This included (1) studies to collect general information on user needs and preferences [47-50,63,84,88,93,104-106,112] or (2) studies focusing on factors for implementation [60,69,85,111]. Other studies were oriented toward actual implementation and (3) described pilot-testing [90] or (4) the overall design process [65,76,77,114]. Further studies were based on refining the existing content such as (5) studies that focused on the redesign of EHRs or prototypes [57,62,66] or (6) studies that included system improvement and further development [57,91,95,99]. In addition, studies have examined implemented EHRs, which were (7) studies in terms of overall satisfaction or acceptance [51,64,67,68,79-81,86,96,107] or (8) studies focusing on terms of usability or system performance [52-54,58,59,61,​70-75,78,87,89,92,94,97,98,100-103].

The included studies covered different forms of EHRs and development, which were divided into 4 different categories:

  1. Information needs for subsequent programing of EHRs were the area of research of 4 studies [47-50]. In these studies, the authors addressed the information needs in EHRs from the perspective of clinicians. For example, Acharya et al [47,51] gathered information needs for oral health information for an EHR, and Ellsworth et al [49] surveyed the information needs for a neonatal intensive care EHR.
  2. In total, 5 studies reported on prototypes or mock-ups [52-56]. For example, Belden et al [57] addressed a clinical note prototype, and Horsky et al [54] developed a prototype of an EHR that allowed clinicians to complete a summary for outpatient visits.
  3. Entire EHRs, including all submodules were addressed in 30 studies [56,58-86]. For example, Nolan et al [73] examined information use and workflow patterns for an EHR in an ICU.
  4. Individual modules of an EHR were addressed in 31 studies [51,57,70,87-113]. These studies focused on individual modules of an EHR rather than an entire EHR. In this category, for example, Aakre et al [87] focused on a module for the automatic calculation of a sequential organ failure assessment calculator in sepsis detection, and Ahluwalia et al [88] focused on dyspnea assessment for palliative care.
Setting

The included studies were conducted in the context of different health care settings. A total of 25 studies were conducted in an unspecified hospital setting [60,62,65,67,70-72,75-78,80-85,​89,91,92,99,100,111,114]. Furthermore, 8 studies were conducted in the context of ICUs [49,55,56,58,​63,68,73,87,93,115], 6 were conducted in the context of family medicine [52,59,101,102,109,110], 6 were conducted in primary care hospitals [53,61,97,98,104,105], 5 were conducted in outpatient or clinic settings [54,57,74,95,108], and 3 were conducted in tertiary hospitals [50,86,107]. Two studies each were conducted in a dental clinic [47,51], in a palliative care setting [88,96], in emergency departments [90,103], in the gynecological and antenatal settings [64,66], and in the context of mental health or psychiatry settings [48,69]. One study each was conducted in the setting of home care [94], older adult care [79], community health [106], cancer centers [112], and childcare [113].

Study Participants

The participants in the included studies comprised a total of 15 different professions (Table 1). Physicians were involved in 76% (53/70) of the studies, whereas nurses were involved in 40% (28/70) of the studies. Pharmacists were involved in 10% (7/70) of the studies, physiotherapists were involved in 6% (4/70) of the studies, social workers were involved in 4% (3/70) of the studies, and medical assistants were involved in 3% (2/70) of the studies. In 16% (11/70) of the studies, the user groups were not specified. Demographic characteristics such as age, sex, and education were described in 21 studies [53,57,67-70,73,75,79,81,84,91-93,96,98,103,105,106,111,113,114]. Moreover, in 3 studies, the authors provided a brief description of demographic characteristics [64,74,97]. For example, one study provided a description of the demographic characteristics as follows: “The sample consisted of 21 female participants and 9 male participants, with a proportion of 70% female and 30% male” [64]. The remaining studies did not describe the demographic characteristics of the participants.

In 10% (7/70) of the included studies, participants received financial compensation for taking part in the study. The following amounts were paid to the participants: US $100 gift card—25 to 45 minutes [92], US $100 gift card [68], US $50 per hour [75], US $100 per 2 hours [76], €40 (US $43) per hour [96], and US $100 per hour [56]. In one study, a US $25 gift card for a restaurant was offered [59].

In addition to the characteristics of the participants, the number of participants in each study was divided into 7 categories (Table 2).

Table 1. User groups included in the studies (n=70).
User groupStudies that included this user group, n (%)Studies
Physicians52 (74)[48-50,52-60,62-71,73,75,76,78,79,82,83,86,88,89,91-93,95-97,100-105,107-110,112-115]
Nurses29 (41)[48-50,60,64,65,69,72,74,78-81,84,85,88,93,94,96,99,100,102,103,105-107,111,113]
Pharmacists7 (10)[65,78,79,82,85,90,91]
Physiotherapists4 (6)[60,78,79,97]
Social workers3 (4)[79,85,93]
Medical assistants2 (3)[74,106]
Psychologists1 (1)[85]
Physician assistants1 (1)[106]
Managers1 (1)[85]
Medical office assistants1 (1)[98]
Midwives1 (1)[66]
Community health agents (CHA)1 (1)[48]
Primary care providers1 (1)[92]
Medical secretaries1 (1)[60]
IT departments, hospital’s IT1 (1)[60]
Not specified11 (16)[47,51,57,61,62,77,78,85,87,105,113]
Table 2. Number of participants per study (n=70).
Range for the number of participantsStudies, n (%)Studies
1-1015 (21)[54-56,63,77,82,83,89,95,97,98,100,108,109,115]
11-2019 (27)[52,53,58-61,66,74-76,87,88,92,96,99,103,104,110]
21-3012 (17)[49,57,64,67-69,72,73,90,102,107,114]
31-405 (7)[71,79,84,94,105]
41-505 (7)[48,65,80,91,101]
51-1005 (7)[51,78,106,111,113]
>1009 (13)[47,50,62,70,81,85,86,93,112]
Methods and Materials

First, the basic methodology of the studies was examined. Overall, 36% (25/70) of the studies used mixed methods design [54,55,57,60,61,66,72,75,78,79,81,83,84,90,92,94,95,98,100,​102,103,105,108,109,111,113]. Moreover, 31% (22/70) of the studies used a qualitative design [49,50,52,56,59,62,64,​69-71,73,74,77,86,87,91,96,99,104,107,114,115], whereas 31% (22/70) of the studies used a quantitative design [47,48,​51,53,58,63,65,67,68,76,80,82,85,88,89,93,97,99,101,106,110,112].

The wide variance of terminology in relation to the involvement of users in technology development already mentioned at the beginning is also reflected in the articles included. The terminology here describes the literal naming of the method by the authors of the studies themselves, regardless of how it was conducted. The largest proportion of studies (44/70, 63%) did not include a designation of methodology [47,51,52,55-62​,64,66,67,69,71,72,76-86,88,90-93,95,98-101,104,106,107,110,​111,115]. For example, in 23% (16/70) of the studies, the authors of the respective manuscripts described the methodological approach as “user-centered design” [54,57,68,70,73-75,79,87,93,97,102,103,108,109,114]. In one of the studies, it was only given as a keyword and not in the manuscript [97]. In terms of frequency, the following terms were used: “participatory design” in 6% (4/70) of the studies [48,63,65,96], “co-design” in 3% (2/70) of the studies [94,105], and “iterative rapid design involving providers” [53], “end-user design” [49], “multidisciplinary design” [61] and “human-centered design” [113] in 1 study each.

The frequency with which users were involved in the development was examined. Involvement counts as renewed involvement if it takes place at a later point and contributes to the further development of the technology (for example, several surveys for the iterative refinement of a prototype). In 57% (40/70) of the studies, users were included once [47,49-51,55,56,58,59,69-73,75-77,80,82-84,88,91-93,95-97,​99-101,103-106,108-111,114,115]. An involvement of users at 2 points was investigated in 24% (17/70) of the studies. Moreover, 9% (6/70) of the studies reported 3 times of user involvement, 4% (3/70) of the studies reported 4 times of involvement, 23% (2/70) of the studies reported 6 times of involvement, 1% (1/69) study reported 5 times of involvement, and 1% (1/69) study reported 9 times of involvement.

In some of the studies, a foundation of the study was provided before the actual (further) development of the EHR. This included, for example, literature reviews [47,61,63,68,90], pilot-testing of the design [52], pilot-testing of the survey [81] or interview guide [47,51,68], a review of 12 different EHRs [57] as well as training with the software in advance [76,77,83,103,107], and the presentation of learning videos [91].

A common method of data collection and involvement of health care professionals was to test a prototype as a walkthrough using think-aloud technique [52,55,56,58,59,71,74-77,83,89,​90,92,93,95-100,103-105,108-110,113-115]. As part of the walkthrough methodology, various programs (eg, Morae) have been used to record audio or screen displays, mouse clicks, and keyboard [52,54,74-76,83,92,94,97,100,103,109,113,114]. Eye-tracking software (eg, Tobii T120 eye tracker) was used [52,59,71,75]. A related method, the near-live testing, was used in one study [89].

Another common method used were the questionnaires. In addition to individually created surveys [47,49,50,60,64,66,70,78,79,83,86,101-103,106,107,113], various existing questionnaires were used (Table 3).

Some of the studies used web-based questionnaire tools (eg, Survey Monkey) [47,50,61,69,70,86,113].

Individual semistructured interviews [48,53-55,60,63,65,67,​68,70,72,79,80,82-85,87,88,90,93,94,96-98,102,104-106,110,111,113,114] and group interviews and focus group discussions [48,60,62,63,67] were conducted. In some of the studies, design workshops were held with various users [53,57,65,90,110].

One method that was often combined with interviews was observation. This involved observing health care professionals as they used an EHR to conduct documentation [60,63,66,73,79,85,87,90,94,102,113,114]. This includes observations in both a clinical and a study setting.

The use of mock-ups was another common method in the studies. This contained paper prototypes [57,90,94] and web-based prototypes using different prototyping tools (eg, HipMunk) [57,87,90,93,94,105,113,114].

Table 3. Questionnaires used in the studies.
QuestionnaireFocus of the questionnaireStudies using the questionnaire
Baylor EHRa user experience surveyMeasuring user experience following EHR implementation[69]
Canada Health Infoway System and Use Assessment SurveyMeasuring user adoption and use as well as information and system quality[81]
Computer Systems Usability QuestionnaireMeasuring satisfaction of users with computer system usability[87]
Information System Use InstrumentMeasuring nurses’ information systems use[81]
Keystroke-level model GOMS Predict or estimate the time for completing a task in software[59]
Nasa Task Load Index Measuring perceived workload[52,58,59,92,113]
Physician Documentation Quality Instrument-9 Assessing the quality of physician electronic documentation[75]
Post-Study System Usability Questionnaire Measuring the perceived satisfaction[100]
Questionnaire for user interaction satisfaction (short form)Measuring the subjective satisfaction with the human-computer interface[94]
Single Ease QuestionAssessing the difficulty of a task[74,90]
System Usability ScaleMeasuring the usability[52,55,70,72,75,77,78,94,100,103,108]
Technology Acceptance Model QuestionnaireMeasuring likelihood of technology acceptance[61]
Usability AssessmentMeasuring usability[72]
Workflow Integration Survey Measuring workflow integration[81]

aEHR: electronic health record.

Less frequently used methods include document analysis [48,85] and extraction of routine data from the EHR for analysis [60,61,70,74,111,113].

An overview of the respective methods by study aim is available in the metadata table in Multimedia Appendix 1.

Frameworks, Theories, and Guidelines

In the following studies, frameworks are understood as approaches that frame the entire research project and influence the choice of methods and their structure. In these studies, the updated DeLone and McLean framework [127] for evaluating information systems success was used once [60]. The design science framework [128] was used once to develop prototype dashboards [94]. In the same study, the tasks, users, representations, and functions framework [129] was used to structure the usability evaluation [94]. Falah et al [64] used the plan-do-study-act cycle [130] to facilitate the implementation process. In addition, Dziadzko et al [86] used the define-measure-analyze-improve-control quality measurement [131] for implementation measurement. The social science approach of lightweight ethnography [132] was used to design the study by Chruscicki et al [90]. Owing to limited time, the predesigned sample, and particular research questions, one study [67] used framework analysis [133] as a framework. Sockolow et al [79] used the health information technology research-based evaluation framework [134] as a framework to design their study and as a theory to merge qualitative and quantitative data. In addition, in 7% (5/70) of the studies, the authors referred to ISO 9241-210 in their theoretical background [59,70,75,113,114]. However, ISO 9241-210 was not used as a theoretical framework in any of the incident studies.

Theories are understood to be those approaches that were used exclusively within the data evaluation or specific analytic method but did not influence the choice of methods for the entire study as a whole. Kernebeck et al [96] used the Unified Theory of Acceptance and Use of Technology [135] to evaluate think-aloud sessions. Wawrzyniak et al [82] used the critical incident technique by Flanagan [136] to design interview protocols. The interview guide in another study [105] was based on the diffusion of innovations theory [137] and complementary ones. The human factors model Systems Engineering Initiative for Patient Safety 2.0 [138] was used by Cohen et al [106] for data collection. The Cognitive Load Theory [139] was described as important and used for evaluation by Curran et al [92]. In addition, the attention capacity model [140], which focuses on mental effort, was used as a theoretical background by Mosaly et al [71]. The technology acceptance model [141] was used to design a questionnaire [113].

In addition, design guidelines were mentioned that influenced the basic logic of the EHRs design. This includes the ergonomics of activity [142] mentioned in one study [48], the suggested time and motion procedures [143] in another study [73], and usability heuristics by Nielsen [144] in 1 study [74]. Another framework used was the spiral model for software development [145] in addition to the EHR system user interface framework of the Veterans Affairs Computerized Patient Record System [146] in 1 study [95]. The data-knowledge-information-wisdom framework [147] was named as an important component of informatics in nursing by Nation et al [72].

Competencies of Researchers

In addition, all studies were screened for the description of the competencies of the researchers who conducted the studies. A distinction can be made between competences related to software knowledge (eg, usability experience and software programming) or competences related to knowledge of the context of use (eg, previous experience of working with EHRs in clinical settings) or methodological skills in the area of data collection (qualitative interviews, surveys). In 30% (21/70) of the studies, the authors briefly described the competencies of the researchers [49,52,53,60,63,68,70,74,75,82,88,93,97,99,​103,104,106,109,110,112,113]. Examples of these descriptions were that the researchers described themselves as “experienced in qualitative research” [106] and “the research team included three academic researchers and two clinical nurses” [99].


This review aimed to provide an overview of the existing methods of user involvement in the literature for developing and evaluating EHRs.

Principal Findings

The review had four objectives: (1) to conduct a systematic search of the published literature for studies focusing on user involvement in the development of EHRs, (2) to present the characteristics and range of methods used in the identified manuscripts, (3) to explore the reported challenges and limitations of the methods, and (4) to make recommendations for further developing the approach and improving the consistency with which they are conducted and reported. Therefore, the main focus of the review was to examine in which settings which participants were involved with which methods and materials and which frameworks were used. Furthermore, the frequency and design of the development and an overview of the competences of the respective researchers involved in the development were examined. To the best of our knowledge, this is the first review to describe the methodological aspects for involving health care professionals in the development of EHRs.

The characteristics of EHRs addressed in the included studies covered a variety of different aspects. On the one hand, a large number of studies addressed a comprehensive EHR, whereas on the other hand, many studies addressed only individual modules of an EHR. The wide range of characteristics in this review was largely because of the broad inclusion criteria, which were designed to provide a comprehensive picture of the methodological aspects of professional involvement in the development of EHRs. This leads to a better description of the complex field of EHRs and their methodological aspects. However, this sometimes makes it difficult to compare the interventions. In terms of setting, it was found that most of the included studies were conducted in general hospitals and ICUs. The authors suggest analyzing studies in a specific setting and how the participants are involved there in the future to cover the methodological aspects, such as in ICU or in palliative care. Future studies could also focus on the individual modules of an EHR, for example, medication modules.

In terms of participants, the studies mainly included physicians and nurses. In terms of multidisciplinary care, it would be desirable for all health care professionals (including physiotherapists, occupational therapists, speech therapists, social workers, nursing assistants, etc) to record their activities and observations in joint documentation and to be able to view them mutually. In addition, sharing EHRs can facilitate communication [148] (eg, by sending messages within a program and assigning tasks). With this in mind, it is surprising that only these 2 professional groups were so intensively involved. It would be desirable for further studies to include all professional groups and to design EHRs to meet their needs. However, most of the studies did not specify which health care professionals were involved in the development. This was partly because of an imprecise naming of the participants and partly because of the lack of naming. Future studies should specifically describe the demographic characteristics of the participants, which may lead to a better assessment of the results [33,36]. This is important because demographic variables have a strong influence on the acceptance of EHRs and the level of competence in using EHRs and digital health technologies in general [149,150]. Therefore, it is recommended that study investigators collect key demographic variables from participants and present them in tabular form to improve the interpretation of the results.

The methodology of the studies was balanced between qualitative, quantitative, and mixed methods. However, in 63% (44/70) of the studies, no terminology was used to describe the design of user involvement in more detail. This adds to the imprecision of the presentation in terms of the level of involvement and a qualitative assessment of the methodology. Although 23% (16/70) of the included studies that mentioned user-centered design as an approach will be examined to see if this was really implemented, most of the studies remain vague about user involvement. This again supports the broad search strategy of the review but also points to qualitative ambiguities of implementation.

The frequency of user involvement varied widely, and in most of the studies (40/70, 57%), users were involved only at one point. This shows that a true participatory design or co-design, as it is called for, is rather rare and fuels the suspicion of sham participation, where user requirements are collected but no iteration is performed to test the fit. Another problem with user involvement in development is that it often occurs at only one point in the development of new technologies. This problem is often referred to as “project-based temporality” in the involvement of users [151]. Therefore, it is recommended that users are involved in the development of new technologies at all stages of development over a longer period [151].

Most of the methods vary widely. Think-aloud approaches were often used to obtain user feedback on an EHR. The continuation of this method, near-live testing, which also increases the likelihood of a good fit, was used in only one study. In the case of questionnaires, individual, nonvalidated questionnaires were frequently used, which reduced the quality of the results. The most commonly used questionnaires were mainly oriented toward usability (System Usability Scale) in 16% (11/70) of the studies and the cognitive load (Nasa Task Load Index) in 7% (5/70) of the studies. Cognitive load refers to the amount of mental effort required by a person to perform a specific task and the associated required capacity of the human working memory [152]. It is imperative to consider cognitive load in the development of EHRs and in the implementation of appropriate solutions in clinical practice, as cognitive load has been linked to the development of burnout and distress [27]. It is recommended that user involvement studies use a mix of methods from the fields of telling, making, and enacting [153]. This emphasizes the impact of user involvement through the exchange of current and future practices and the sharing of needs (telling). For successful user involvement, it is crucial that future users develop something (making), contributing to the existence and design of a new technology. In addition, by involving users, it should be possible to transfer ideas into reality by creating a simulation to test them [153]. Accordingly, different questions require different methods from each field, but a mixed methods approach ensures a diversity of perspectives.

Frameworks, theories, and guidelines were very rarely used. Moreover, the results showed a rather low level of consideration of the theoretical underpinnings to the detriment of the quality of the studies. This is particularly problematic because the use of such frameworks can structure the development and make the replicability of results between different studies comparable. Furthermore, it is problematic that a large number of studies did not refer to theories and models. This would also provide a theoretical basis for the development and make the quality of the results more comprehensible [32]. It is particularly surprising that only 9% (6/70) of the studies referenced to ISO standards. These standardizes the process for the development of new software. It would be useful to refer to these standards in future studies and to highlight the stage of development of the respective technology [154,155].

The researchers’ skills have rarely been documented. A multidisciplinary research and development team should consist of individuals with different skills from different health care disciplines, methodological disciplines, and social disciplines. This allows for optimal design and support of different stakeholders during development and implementation [156]. Similar to the sometimes imprecise description of demographic characteristics of the study participants, the skills of the investigators should be described. Owing to the interdisciplinary nature of technology development research projects, it would be advisable to have multidisciplinary teams and to identify the respective competencies and experience in technology development.

Limitations

To be able to interpret these results, it is necessary to describe several limitations of this study. First, the search strategy was limited by its focus on empirical, scientifically published work. The involvement of health care professionals in the development of EHRs may not always be published in scientific journals. Therefore, this review is a first step on the topic of involving health care professionals in the development of EHRs. In further studies, it might be interesting to include gray literature and databases with a focus on technology-oriented research and engineering (eg, Institute of Electrical and Electronics Engineers). However, the heterogeneity of the quality of the publications must be taken into account. In addition, EHRs are mostly developed by large digital technology companies. It can be assumed that these companies often involve users in the development but do not produce publications or perform actual research. Therefore, it can be assumed that there was publication bias. It would be necessary in the future to survey such large companies on how they involve users in the development of EHRs.

Second, it should be noted that the screening process was limited by the definitions of user involvement, which accordingly shaped both the search terms and the inclusion and exclusion criteria. Although a wide range of terms were used, it cannot be ruled out that individual manuscripts that were coherent in terms of content, and therefore would have led to different results, were not included because of the lack of used terminology.

Third, it should be taken into account that some studies published their results in several manuscripts and did not briefly review the entire development or implementation process. In our evaluation, we were only able to consider the described frequency of user involvement from the information provided in the included manuscripts. However, it is possible that the included manuscripts each report only a subset of the study project; whereas in the overall study project, users were much more frequently involved.

Finally, one of the findings, namely the sometimes-low transparency of reporting, also directly points to a limitation in terms of analysis and conclusions—drawing conclusions on the basis of reporting in manuscripts has limited validity, as there is no way to ensure that the actual methodological considerations and intentions correspond to what was presented in the manuscript.

Further Research

Further research can help to improve the methodological framework for involving health care professionals in the development of EHRs. Particular attention should be paid to the rationale for the methodological choices. It is also crucial to combine different methods from the fields of telling, making, and enacting; involve users at several points in time to avoid sham participation; and strive for maximum user orientation. The growing interest in “design through design research” [157] should be encouraged but with conditions to promote high-quality developments. Little knowledge can be gained from publications with low reporting quality in terms of transferability and quality assessments. It would be useful for studies to report the aspects more precisely. Specific reporting guidelines for reporting the results of technology development studies would be helpful, as is the case for many other types of studies [158]. Process evaluations should be used in a standardized manner to improve study quality.

In addition, in future studies, it would be necessary to examine in more detail the outcomes of the participatory design with users. In this context, questions should be answered regarding the specific outcomes that have been improved by the involvement of users. How these outcomes were measured and how, for example, improvements to the software were evaluated in different iterations should also be analyzed.

Conclusions

Studies involving health care professionals in the development of EHRs have used various approaches. This paper provides an overview of the approaches in different fields of development with the inclusion of diverse users. Often, however, there is no specific approach, framework, or theory underlying the procedure and there is missing or inaccurate information in the reporting.

Acknowledgments

The authors acknowledge support from the Open Access Publication Funds of the Ruhr-Universität Bochum.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Metadata of the included studies.

DOCX File , 65 KB

Multimedia Appendix 2

Search string.

DOCX File , 27 KB

  1. Gonçalves AS, Bertram N, Amelung V. European Scorecard zum Stand der Implementierung der elektronischen Patientenakte auf nationaler Ebene. Stiftung Muench. 2018. URL: https://www.stiftung-muench.org/wp-content/uploads/2018/09/Scorecard-final.pdf [accessed 2022-07-21]
  2. Al-Aswad AM, Brownsell S, Palmer R, Nichol JP. A review paper of the current status of electronic health records adoption worldwide: the gap between developed and developing countries. J Health Inform Dev Ctries. 2013;7(2):153-164. [FREE Full text]
  3. Campanella P, Lovato E, Marone C, Fallacara L, Mancuso A, Ricciardi W, et al. The impact of electronic health records on healthcare quality: a systematic review and meta-analysis. Eur J Public Health. Feb 2016;26(1):60-64. [CrossRef] [Medline]
  4. Uslu A, Stausberg J. Value of the electronic medical record for hospital care: update from the literature. J Med Internet Res. Dec 23, 2021;23(12):e26323. [FREE Full text] [CrossRef] [Medline]
  5. Atasoy H, Greenwood BN, McCullough JS. The digitization of patient care: a review of the effects of electronic health records on health care quality and utilization. Annu Rev Public Health. Apr 01, 2019;40:487-500. [CrossRef] [Medline]
  6. Roehrs A, da Costa CA, Righi RD, de Oliveira KS. Personal health records: a systematic literature review. J Med Internet Res. Jun 06, 2017;19(1):e13. [FREE Full text] [CrossRef] [Medline]
  7. Rouleau G, Gagnon MP, Côté J, Payne-Gagnon J, Hudson E, Dubois CA. Impact of information and communication technologies on nursing care: results of an overview of systematic reviews. J Med Internet Res. Apr 25, 2017;19(4):e122. [FREE Full text] [CrossRef] [Medline]
  8. Tsai CH, Eghdam A, Davoody N, Wright G, Flowerday S, Koch S. Effects of electronic health record implementation and barriers to adoption and use: a scoping review and qualitative analysis of the content. Life (Basel). Dec 04, 2020;10(12):327. [FREE Full text] [CrossRef] [Medline]
  9. Lipscombe LL, Hwee J, Webster L, Shah BR, Booth GL, Tu K. Identifying diabetes cases from administrative data: a population-based validation study. BMC Health Serv Res. May 02, 2018;18(1):316. [FREE Full text] [CrossRef] [Medline]
  10. Ebad SA. Healthcare software design and implementation—a project failure case. Softw Pract Exp. Feb 24, 2020;50(7):1258-1276. [FREE Full text] [CrossRef]
  11. Shea CM, Halladay JR, Reed D, Daaleman TP. Integrating a health-related-quality-of-life module within electronic health records: a comparative case study assessing value added. BMC Health Serv Res. Mar 19, 2012;12:67. [FREE Full text] [CrossRef] [Medline]
  12. Jorzig A, Sarangi F. Digitalisierung im Gesundheitswesen: Ein kompakter Streifzug durch Recht, Technik und Ethik. Berlin, Germany. Springer; May 2020.
  13. Evans RS. Electronic health records: then, now, and in the future. Yearb Med Inform. May 20, 2016;Suppl 1(Suppl 1):S48-S61. [FREE Full text] [CrossRef] [Medline]
  14. Schneider EC, Ridgely MS, Meeker D, Hunter LE, Khodyakov D, Rudin RS. Promoting patient safety through effective health information technology risk management. Rand Health Q. Dec 30, 2014;4(3):7. [FREE Full text] [Medline]
  15. Jones SS, Rudin RS, Perry T, Shekelle PG. Health information technology: an updated systematic review with a focus on meaningful use. Ann Intern Med. Jan 07, 2014;160(1):48-54. [FREE Full text] [CrossRef] [Medline]
  16. Sutton RT, Pincock D, Baumgart DC, Sadowski DC, Fedorak RN, Kroeker KI. An overview of clinical decision support systems: benefits, risks, and strategies for success. NPJ Digit Med. Feb 06, 2020;3:17. [FREE Full text] [CrossRef] [Medline]
  17. Wisner K, Chesla CA, Spetz J, Lyndon A. Managing the tension between caring and charting: labor and delivery nurses' experiences of the electronic health record. Res Nurs Health. Oct 2021;44(5):822-832. [CrossRef] [Medline]
  18. Ferrell BR, Twaddle ML, Melnick A, Meier DE. National consensus project clinical practice guidelines for quality palliative care guidelines, 4th edition. J Palliat Med. Dec 2018;21(12):1684-1689. [CrossRef] [Medline]
  19. Senathirajah Y, Kaufman DR, Cato KD, Borycki EM, Fawcett JA, Kushniruk AW. Characterizing and visualizing display and task fragmentation in the electronic health record: mixed methods design. JMIR Hum Factors. Oct 21, 2020;7(4):e18484. [FREE Full text] [CrossRef] [Medline]
  20. Vanderhook S, Abraham J. Unintended consequences of EHR systems: a narrative review. Proc Int Symp Hum Factors Ergon health care. May 15, 2017;6(1):218-225. [FREE Full text] [CrossRef]
  21. Kaihlanen AM, Gluschkoff K, Hyppönen H, Kaipio J, Puttonen S, Vehko T, et al. The associations of electronic health record usability and user age with stress and cognitive failures among Finnish registered nurses: cross-sectional study. JMIR Med Inform. Nov 18, 2020;8(11):e23623. [FREE Full text] [CrossRef] [Medline]
  22. Wisner K, Lyndon A, Chesla CA. The electronic health record's impact on nurses' cognitive work: an integrative review. Int J Nurs Stud. Jun 2019;94:74-84. [CrossRef] [Medline]
  23. Li C, Parpia C, Sriharan A, Keefe DT. Electronic medical record-related burnout in healthcare providers: a scoping review of outcomes and interventions. BMJ Open. Aug 19, 2022;12(8):e060865. [FREE Full text] [CrossRef] [Medline]
  24. Melnick ER, West CP, Nath B, Cipriano PF, Peterson C, Satele DV, et al. The association between perceived electronic health record usability and professional burnout among US nurses. J Am Med Inform Assoc. Jul 30, 2021;28(8):1632-1641. [FREE Full text] [CrossRef] [Medline]
  25. Khairat S, Coleman C, Ottmar P, Jayachander DI, Bice T, Carson SS. Association of electronic health record use with physician fatigue and efficiency. JAMA Netw Open. Jun 01, 2020;3(6):e207385. [FREE Full text] [CrossRef] [Medline]
  26. Melnick ER, Dyrbye LN, Sinsky CA, Trockel M, West CP, Nedelec L, et al. The association between perceived electronic health record usability and professional burnout among US physicians. Mayo Clin Proc. Mar 2020;95(3):476-487. [FREE Full text] [CrossRef] [Medline]
  27. Melnick ER, Harry E, Sinsky CA, Dyrbye LN, Wang H, Trockel MT, et al. Perceived electronic health record usability as a predictor of task load and burnout among US physicians: mediation analysis. J Med Internet Res. Dec 22, 2020;22(12):e23382. [FREE Full text] [CrossRef] [Medline]
  28. Ratwani RM, Savage E, Will A, Fong A, Karavite D, Muthu N, et al. Identifying electronic health record usability and safety challenges in pediatric settings. Health Aff (Millwood). Nov 2018;37(11):1752-1759. [CrossRef] [Medline]
  29. Ross J, Stevenson F, Lau R, Murray E. Factors that influence the implementation of e-health: a systematic review of systematic reviews (an update). Implement Sci. Oct 26, 2016;11(1):146. [FREE Full text] [CrossRef] [Medline]
  30. Kruse CS, Kristof C, Jones B, Mitchell E, Martinez A. Barriers to electronic health record adoption: a systematic literature review. J Med Syst. Dec 2016;40(12):252. [FREE Full text] [CrossRef] [Medline]
  31. Boyd H, McKernon S, Mullin B, Old A. Improving healthcare through the use of co-design. N Z Med J. Jun 29, 2012;125(1357):76-87. [Medline]
  32. Moore G, Wilding H, Gray K, Castle D. Participatory methods to engage health service users in the development of electronic health resources: systematic review. J Particip Med. Feb 22, 2019;11(1):e11474. [FREE Full text] [CrossRef] [Medline]
  33. Vandekerckhove P, de Mul M, Bramer WM, de Bont AA. Generative participatory design methodology to develop electronic health interventions: systematic literature review. J Med Internet Res. Apr 27, 2020;22(4):e13780. [FREE Full text] [CrossRef] [Medline]
  34. Sumner J, Chong LS, Bundele A, Wei Lim Y. Co-designing technology for aging in place: a systematic review. Gerontologist. Sep 13, 2021;61(7):e395-e409. [FREE Full text] [CrossRef] [Medline]
  35. Duque E, Fonseca G, Vieira H, Gontijo G, Ishitani L. A systematic literature review on user centered design and participatory design with older people. In: Proceedings of the 18th Brazilian Symposium on Human Factors in Computing Systems. Presented at: IHC '19; October 22-25, 2019, 2019;1-11; Vitória Espírito Santo, Brazil. URL: https://dl.acm.org/doi/10.1145/3357155.3358471 [CrossRef]
  36. Merkel S, Kucharski A. Participatory design in gerontechnology: a systematic literature review. Gerontologist. Jan 09, 2019;59(1):e16-e25. [CrossRef] [Medline]
  37. Fischer B, Peine A, Östlund B. The importance of user involvement: a systematic review of involving older users in technology design. Gerontologist. Sep 15, 2020;60(7):e513-e523. [FREE Full text] [CrossRef] [Medline]
  38. Panek I, Crumley ET, Ishigami-Doyle Y, Sixsmith J, Kontos P, O?Doherty K, et al. Levels of older adults’ engagement in technology research, design and development: a scoping review. Innov Aging. Jul 2017;1(suppl_1):393. [FREE Full text] [CrossRef]
  39. Clemensen J, Rothmann MJ, Smith AC, Caffery LJ, Danbjorg DB. Participatory design methods in telemedicine research. J Telemed Telecare. Oct 2017;23(9):780-785. [CrossRef] [Medline]
  40. Joddrell P, Astell AJ. Studies involving people with dementia and touchscreen technology: a literature review. JMIR Rehabil Assist Technol. Nov 04, 2016;3(2):e10. [FREE Full text] [CrossRef] [Medline]
  41. Dekker MR, Williams AD. The use of user-centered participatory design in serious games for anxiety and depression. Games Health J. Dec 2017;6(6):327-333. [CrossRef] [Medline]
  42. Eyles H, Jull A, Dobson R, Firestone R, Whittaker R, Te Morenga L, et al. Co-design of mHealth delivered interventions: a systematic review to assess key methods and processes. Curr Nutr Rep. Jul 04, 2016;5(3):160-167. [FREE Full text] [CrossRef]
  43. DeSmet A, Thompson D, Baranowski T, Palmeira A, Verloigne M, De Bourdeaudhuij I. Is participatory design associated with the effectiveness of serious digital games for healthy lifestyle promotion? A meta-analysis. J Med Internet Res. Apr 29, 2016;18(4):e94. [FREE Full text] [CrossRef] [Medline]
  44. Göttgens I, Oertelt-Prigione S. The application of human-centered design approaches in health research and innovation: a narrative review of current practices. JMIR Mhealth Uhealth. Dec 06, 2021;9(12):e28102. [FREE Full text] [CrossRef] [Medline]
  45. Simonsen J, Robertson T. Routledge International Handbook of Participatory Design. New York, NY, USA. Routledge; 2013.
  46. Vargas C, Whelan J, Brimblecombe J, Allender S. Co-creation, co-design, co-production for public health - a perspective on definition and distinctions. Public Health Res Pract. Jun 15, 2022;32(2):3222211. [FREE Full text] [CrossRef] [Medline]
  47. Acharya A, Mahnke A, Chyou PH, Rottscheit C, Starren JB. Medical providers' dental information needs: a baseline survey. Stud Health Technol Inform. 2011;169:387-391. [FREE Full text] [Medline]
  48. do Carmo Alonso CM, de Lima AN, Oggioni BM, Teixeira MR, Oliveira EP, Couto MC, et al. Contributions of activity ergonomics to the design of an electronic health record to support collaborative mental care of children and youth: preliminary results. Work. 2020;65(1):187-194. [CrossRef] [Medline]
  49. Ellsworth MA, Lang TR, Pickering BW, Herasevich V. Clinical data needs in the neonatal intensive care unit electronic medical record. BMC Med Inform Decis Mak. Oct 24, 2014;14:92. [FREE Full text] [CrossRef] [Medline]
  50. Herasevich V, Ellsworth MA, Hebl JR, Brown MJ, Pickering BW. Information needs for the OR and PACU electronic medical record. Appl Clin Inform. Jul 16, 2014;5(3):630-641. [FREE Full text] [CrossRef] [Medline]
  51. Acharya A, Shimpi N, Mahnke A, Mathias R, Ye Z. Medical care providers' perspectives on dental information needs in electronic health records. J Am Dent Assoc. May 2017;148(5):328-337. [CrossRef] [Medline]
  52. Belden JL, Koopman RJ, Patil SJ, Lowrance NJ, Petroski GF, Smith JB. Dynamic electronic health record note prototype: seeing more by showing less. J Am Board Fam Med. Nov 2017;30(6):691-700. [FREE Full text] [CrossRef] [Medline]
  53. Mishuris RG, Yoder J, Wilson D, Mann D. Integrating data from an online diabetes prevention program into an electronic health record and clinical workflow, a design phase usability study. BMC Med Inform Decis Mak. Jul 11, 2016;16:88. [FREE Full text] [CrossRef] [Medline]
  54. Horsky J, Ramelson HZ. Development of a cognitive framework of patient record summary review in the formative phase of user-centered design. J Biomed Inform. Dec 2016;64:147-157. [FREE Full text] [CrossRef] [Medline]
  55. King AJ, Cooper GF, Hochheiser H, Clermont G, Visweswaran S. Development and preliminary evaluation of a prototype of a learning electronic medical record system. AMIA Annu Symp Proc. Nov 05, 2015;2015:1967-1975. [FREE Full text] [Medline]
  56. King AJ, Cooper GF, Hochheiser H, Clermont G, Hauskrecht M, Visweswaran S. Using machine learning to predict the information seeking behavior of clinicians using an electronic medical record system. AMIA Annu Symp Proc. Dec 05, 2018;2018:673-682. [FREE Full text] [Medline]
  57. Belden JL, Wegier P, Patel J, Hutson A, Plaisant C, Moore JL, et al. Designing a medication timeline for patients and physicians. J Am Med Inform Assoc. Feb 01, 2019;26(2):95-105. [FREE Full text] [CrossRef] [Medline]
  58. Ahmed A, Chandra S, Herasevich V, Gajic O, Pickering BW. The effect of two different electronic health record user interfaces on intensive care provider task load, errors of cognition, and performance. Crit Care Med. Jul 2011;39(7):1626-1634. [CrossRef] [Medline]
  59. Al Ghalayini M, Antoun J, Moacdieh NM. Too much or too little? Investigating the usability of high and low data displays of the same electronic medical record. Health Informatics J. Mar 2020;26(1):88-103. [FREE Full text] [CrossRef] [Medline]
  60. Bossen C, Jensen LG, Udsen FW. Evaluation of a comprehensive EHR based on the DeLone and McLean model for IS success: approach, results, and success factors. Int J Med Inform. Oct 2013;82(10):940-953. [CrossRef] [Medline]
  61. Briggs BW, Carter-Templeton H. Electronic health record customization: a quality improvement project. Online J Nurs Inform. Jan 2014;18(3) [FREE Full text]
  62. Brokel J, Ochylski S, Kramer J. Re-engineering workflows: changing the life cycle of an electronic health record system. J Healthc Eng. Sep 2011;2(3):303-320. [FREE Full text] [CrossRef]
  63. Calzoni L, Clermont G, Cooper GF, Visweswaran S, Hochheiser H. Graphical presentations of clinical data in a learning electronic medical record. Appl Clin Inform. Aug 2020;11(4):680-691. [FREE Full text] [CrossRef] [Medline]
  64. Falah J, Alfalah S, Halawani S, Abulebbeh M, Muhaidat N. EMR for obstetric emergency department and labour ward in Jordan University Hospital. J Theor Appl Inf Technol. Jul 15, 2020;98(13):2253. [FREE Full text]
  65. Grenha Teixeira J, Pinho NF, Patrício L. Bringing service design to the development of health information systems: the case of the Portuguese national electronic health record. Int J Med Inform. Dec 2019;132:103942. [CrossRef] [Medline]
  66. Helou S, Abou-Khalil V, Yamamoto G, Kondoh E, Tamura H, Hiragi S, et al. Understanding the situated roles of electronic medical record systems to enable redesign: mixed methods study. JMIR Hum Factors. Jul 09, 2019;6(3):e13812. [FREE Full text] [CrossRef] [Medline]
  67. Hernández-Ávila JE, Palacio-Mejía LS, Lara-Esqueda A, Silvestre E, Agudelo-Botero M, Diana ML, et al. Assessing the process of designing and implementing electronic health records in a statewide public health system: the case of Colima, Mexico. J Am Med Inform Assoc. Mar 2013;20(2):238-244. [FREE Full text] [CrossRef] [Medline]
  68. Khairat S, Coleman C, Teal R, Rezk S, Rand V, Bice T, et al. Physician experiences of screen-level features in a prominent electronic health record: design recommendations from a qualitative study. Health Informatics J. Jan 2021;27(1):1460458221997914. [FREE Full text] [CrossRef] [Medline]
  69. Lipford K, Jones S, Johnson K. Needs assessment of an electronic health record at an inpatient psychiatric hospital. Online J Nurs Inform. Feb 2017;21(1) [FREE Full text] [CrossRef]
  70. Luna DR, Rizzato Lede DA, Rubin L, Otero CM, Ortiz JM, García MG, et al. User-centered design improves the usability of drug-drug interaction alerts: a validation study in the real scenario. Stud Health Technol Inform. 2017;245:1085-1089. [Medline]
  71. Mosaly PR, Guo H, Mazur L. Toward better understanding of task difficulty during physicians’ interaction with Electronic Health Record System (EHRs). Int J Hum Comput Interact. Feb 20, 2019;35(20):1883-1891. [FREE Full text] [CrossRef]
  72. Nation J, Wangia-Anderson V. Applying the data-knowledge-information-wisdom framework to a usability evaluation of electronic health record system for nursing professionals. Online J Nurs Inform. 2019;23(1) [FREE Full text]
  73. Nolan ME, Siwani R, Helmi H, Pickering BW, Moreno-Franco P, Herasevich V. Health IT usability focus section: data use and navigation patterns among medical ICU clinicians during electronic chart review. Appl Clin Inform. Oct 2017;8(4):1117-1126. [FREE Full text] [CrossRef] [Medline]
  74. Pierce RP, Eskridge BR, Rehard L, Ross B, Day MA, Belden JL. The effect of electronic health record usability redesign on annual screening rates in an ambulatory setting. Appl Clin Inform. Aug 2020;11(4):580-588. [FREE Full text] [CrossRef] [Medline]
  75. Rizvi RF, Marquard JL, Seywerd MA, Adam TJ, Elison JT, Hultman GM, et al. Usability evaluation of an EHR's clinical notes interface from the perspective of attending and resident physicians: an exploratory study. Stud Health Technol Inform. 2017;245:1128-1132. [Medline]
  76. Senathirajah Y, Kaufman D, Bakken S. The clinician in the driver's seat: part 2 - intelligent uses of space in a drag/drop user-composable electronic health record. J Biomed Inform. Dec 2014;52:177-188. [FREE Full text] [CrossRef] [Medline]
  77. Shah KG, Slough TL, Yeh PT, Gombwa S, Kiromera A, Oden ZM, et al. Novel open-source electronic medical records system for palliative care in low-resource settings. BMC Palliat Care. Aug 14, 2013;12(1):31. [FREE Full text] [CrossRef] [Medline]
  78. Snowden A, Kolb H. Two years of unintended consequences: introducing an electronic health record system in a hospice in Scotland. J Clin Nurs. May 2017;26(9-10):1414-1427. [FREE Full text] [CrossRef] [Medline]
  79. Sockolow PS, Bowles KH, Lehmann HP, Abbott PA, Weiner JP. Community-based, interdisciplinary geriatric care team satisfaction with an electronic health record: a multimethod study. Comput Inform Nurs. Jun 2012;30(6):300-311. [CrossRef] [Medline]
  80. Stevenson JE, Nilsson G. Nurses' perceptions of an electronic patient record from a patient safety perspective: a qualitative study. J Adv Nurs. Mar 2012;68(3):667-676. [CrossRef] [Medline]
  81. Strudwick G, McGillis Hall L, Nagle L, Trbovich P. Acute care nurses' perceptions of electronic health record use: a mixed method study. Nurs Open. May 07, 2018;5(4):491-500. [FREE Full text] [CrossRef] [Medline]
  82. Wawrzyniak C, Marcilly R, Baclet N, Hansske A, Pelayo S. EHR usage problems: a preliminary study. Stud Health Technol Inform. 2019;257:484-488. [Medline]
  83. Zhu X, Cimino JJ. Clinicians' evaluation of computer-assisted medication summarization of electronic medical records. Comput Biol Med. Apr 2015;59:221-231. [FREE Full text] [CrossRef] [Medline]
  84. Carrington JM, Effken JA. Strengths and limitations of the electronic health record for documenting clinical events. Comput Inform Nurs. Jun 2011;29(6):360-367. [CrossRef] [Medline]
  85. Cresswell KM, Worth A, Sheikh A. Integration of a nationally procured electronic health record system into user work practices. BMC Med Inform Decis Mak. Mar 08, 2012;12:15. [FREE Full text] [CrossRef] [Medline]
  86. Dziadzko MA, Herasevich V, Sen A, Pickering BW, Knight AM, Moreno Franco P. User perception and experience of the introduction of a novel critical care patient viewer in the ICU setting. Int J Med Inform. Apr 2016;88:86-91. [CrossRef] [Medline]
  87. Aakre CA, Kitson JE, Li M, Herasevich V. Iterative user interface design for automated sequential organ failure assessment score calculator in sepsis detection. JMIR Hum Factors. May 18, 2017;4(2):e14. [FREE Full text] [CrossRef] [Medline]
  88. Ahluwalia SC, Leos RL, Goebel JR, Asch SM, Lorenz KA. Provider approaches to palliative dyspnea assessment: implications for informatics-based clinical tools. Am J Hosp Palliat Care. May 2013;30(3):231-238. [CrossRef] [Medline]
  89. Chrimes D, Kitos NR, Kushniruk A, Mann DM. Usability testing of Avoiding Diabetes Thru Action Plan Targeting (ADAPT) decision support for integrating care-based counseling of pre-diabetes in an electronic health record. Int J Med Inform. Sep 2014;83(9):636-647. [FREE Full text] [CrossRef] [Medline]
  90. Chruscicki A, Badke K, Peddie D, Small S, Balka E, Hohl CM. Pilot-testing an adverse drug event reporting form prior to its implementation in an electronic health record. Springerplus. Oct 11, 2016;5(1):1764. [FREE Full text] [CrossRef] [Medline]
  91. Cooler J, Ross CA, Robert S, Linder L, Ruhe AM, Philip A. Evaluation and optimization of take-home naloxone in an academic medical center. Ment Health Clin. Mar 01, 2019;9(2):105-109. [FREE Full text] [CrossRef] [Medline]
  92. Curran RL, Kukhareva PV, Taft T, Weir CR, Reese TJ, Nanjo C, et al. Integrated displays to improve chronic disease management in ambulatory care: a SMART on FHIR application informed by mixed-methods user testing. J Am Med Inform Assoc. Aug 01, 2020;27(8):1225-1234. [FREE Full text] [CrossRef] [Medline]
  93. Desai AV, Michael CL, Kuperman GJ, Jordan G, Mittelstaedt H, Epstein AS, et al. A novel patient values tab for the electronic health record: a user-centered design approach. J Med Internet Res. Feb 17, 2021;23(2):e21615. [FREE Full text] [CrossRef] [Medline]
  94. Dowding D, Merrill JA, Barrón Y, Onorato N, Jonas K, Russell D. Usability evaluation of a dashboard for home care nurses. Comput Inform Nurs. Jan 2019;37(1):11-19. [FREE Full text] [CrossRef] [Medline]
  95. Farri O, Rahman A, Monsen KA, Zhang R, Pakhomov SV, Pieczkiewicz DS, et al. Impact of a prototype visualization tool for new information in EHR clinical documents. Appl Clin Inform. Oct 31, 2012;3(4):404-418. [FREE Full text] [CrossRef] [Medline]
  96. Kernebeck S, Busse TS, Jux C, Meyer D, Dreier LA, Zenz D, et al. Participatory design of an electronic medical record for paediatric palliative care: a think-aloud study with nurses and physicians. Children (Basel). Aug 12, 2021;8(8):695. [FREE Full text] [CrossRef] [Medline]
  97. Neudorf B, Giangregorio L, Morita P. Insights from an usability review of an electronic medical record–integrated physical activity counseling tool for primary care. Ergon Des. Jan 2023;31(1):13-22. [FREE Full text] [CrossRef]
  98. Press A, DeStio C, McCullagh L, Kapoor S, Morley J, SBIRT NY-II Team; et al. Usability testing of a national substance use screening tool embedded in electronic health records. JMIR Hum Factors. Jul 08, 2016;3(2):e18. [FREE Full text] [CrossRef] [Medline]
  99. Rogers ML, Sockolow PS, Bowles KH, Hand KE, George J. Use of a human factors approach to uncover informatics needs of nurses in documentation of care. Int J Med Inform. Nov 2013;82(11):1068-1074. [CrossRef] [Medline]
  100. Schall Jr MC, Cullen L, Pennathur P, Chen H, Burrell K, Matthews G. Usability evaluation and implementation of a health information technology dashboard of evidence-based quality indicators. Comput Inform Nurs. Jun 2017;35(6):281-288. [CrossRef] [Medline]
  101. Tazzeo C, Pritchard JM, Papaioannou A, Adachi JD. Promoting osteoporosis best practices: a new electronic medical record tool. J Am Med Dir Assoc. Sep 2020;21(9):1349-1352. [CrossRef] [Medline]
  102. Tran K, Leblanc K, Valentinis A, Kavanagh D, Zahr N, Ivers NM. Evaluating the usability and perceived impact of an electronic medical record toolkit for atrial fibrillation management in primary care: a mixed-methods study incorporating human factors design. JMIR Hum Factors. Feb 17, 2016;3(1):e7. [FREE Full text] [CrossRef] [Medline]
  103. Wang X, Kim TC, Hegde S, Hoffman DJ, Benda NC, Franklin ES, et al. Design and evaluation of an integrated, patient-focused electronic health record display for emergency medicine. Appl Clin Inform. Aug 2019;10(4):693-706. [FREE Full text] [CrossRef] [Medline]
  104. Clarke MA, Steege LM, Moore JL, Koopman RJ, Belden JL, Kim MS. Determining primary care physician information needs to inform ambulatory visit note display. Appl Clin Inform. 2014;05(01):169-190. [FREE Full text] [CrossRef]
  105. Clark RE, Milligan J, Ashe MC, Faulkner G, Canfield C, Funnell L, et al. A patient-oriented approach to the development of a primary care physical activity screen for embedding into electronic medical records. Appl Physiol Nutr Metab. Jun 2021;46(6):589-596. [CrossRef] [Medline]
  106. Cohen DJ, Wyte-Lake T, Dorr DA, Gold R, Holden RJ, Koopman RJ, et al. Unmet information needs of clinical teams delivering care to complex patients and design strategies to address those needs. J Am Med Inform Assoc. May 01, 2020;27(5):690-699. [FREE Full text] [CrossRef] [Medline]
  107. Gascón F, Herrera I, Vázquez C, Jiménez P, Jiménez J, Real C, et al. Electronic health record: design and implementation of a lab test request module. Int J Med Inform. Jun 2013;82(6):514-521. [CrossRef] [Medline]
  108. Hultman G, Marquard J, Arsoniadis E, Mink P, Rizvi R, Ramer T, et al. Usability testing of two ambulatory EHR navigators. Appl Clin Inform. Jun 15, 2016;7(2):502-515. [FREE Full text] [CrossRef] [Medline]
  109. Koopman RJ, Kochendorfer KM, Moore JL, Mehr DR, Wakefield DS, Yadamsuren B, et al. A diabetes dashboard and physician efficiency and accuracy in accessing data needed for high-quality diabetes care. Ann Fam Med. Sep 2011;9(5):398-405. [FREE Full text] [CrossRef] [Medline]
  110. Koopman RJ, Steege LM, Moore JL, Clarke MA, Canfield SM, Kim MS, et al. Physician information needs and Electronic Health Records (EHRs): time to reengineer the clinic note. J Am Board Fam Med. May 2015;28(3):316-323. [FREE Full text] [CrossRef] [Medline]
  111. Oyugi B, Makunja S, Kabuti W, Nyongesa C, Schömburg M, Kibe V, et al. Improving the management of hypertension and diabetes: an implementation evaluation of an electronic medical record system in Nairobi County, Kenya. Int J Med Inform. Sep 2020;141:104220. [CrossRef] [Medline]
  112. Desai AV, Agarwal R, Epstein AS, Kuperman GJ, Michael CL, Mittelstaedt H, et al. Needs and perspectives of cancer center stakeholders for access to patient values in the electronic health record. JCO Oncol Pract. Oct 2021;17(10):e1524-e1536. [FREE Full text] [CrossRef] [Medline]
  113. Thayer JG, Ferro DF, Miller JM, Karavite D, Grundmeier RW, Utidjian L, et al. Human-centered development of an electronic health record-embedded, interactive information visualization in the emergency department using fast healthcare interoperability resources. J Am Med Inform Assoc. Jul 14, 2021;28(7):1401-1410. [FREE Full text] [CrossRef] [Medline]
  114. Luna DR, Rizzato Lede DA, Otero CM, Risk MR, González Bernaldo de Quirós F. User-centered design improves the usability of drug-drug interaction alerts: experimental comparison of interfaces. J Biomed Inform. Feb 2017;66:204-213. [FREE Full text] [CrossRef] [Medline]
  115. King AJ, Cooper GF, Clermont G, Hochheiser H, Hauskrecht M, Sittig DF, et al. Using machine learning to selectively highlight patient information. J Biomed Inform. Dec 2019;100:103327. [FREE Full text] [CrossRef] [Medline]
  116. Peters MD, Godfrey CM, Khalil H, McInerney P, Parker D, Soares CB. Guidance for conducting systematic scoping reviews. Int J Evid Based Healthc. Sep 2015;13(3):141-146. [CrossRef] [Medline]
  117. Peterson J, Pearce PF, Ferguson LA, Langford CA. Understanding scoping reviews: definition, purpose, and process. J Am Assoc Nurse Pract. Jan 2017;29(1):12-16. [CrossRef] [Medline]
  118. Pham MT, Rajić A, Greig JD, Sargeant JM, Papadopoulos A, McEwen SA. A scoping review of scoping reviews: advancing the approach and enhancing the consistency. Res Synth Methods. Dec 2014;5(4):371-385. [FREE Full text] [CrossRef] [Medline]
  119. Li T, Higgins JP, Deeks J. Collecting data. In: Higgins JP, Thomas J, Chandler J, Cumpston M, Li T, Page MJ, et al, editors. Cochrane Handbook for Systematic Reviews of Interventions. 2nd edition. Hoboken, NJ, USA. Wiley-Blackwell; 2019;109-141.
  120. Maggio LA, Larsen K, Thomas A, Costello JA, Artino Jr AR. Scoping reviews in medical education: a scoping review. Med Educ. Jun 2021;55(6):689-700. [FREE Full text] [CrossRef] [Medline]
  121. Brien SE, Lorenzetti DL, Lewis S, Kennedy J, Ghali WA. Overview of a formal scoping review on health system report cards. Implement Sci. Jan 15, 2010;5:2. [FREE Full text] [CrossRef] [Medline]
  122. Iannizzi C, Akl EA, Kahale LA, Dorando E, Mosunmola Aminat A, Barker JM, et al. Methods and guidance on conducting, reporting, publishing and appraising living systematic reviews: a scoping review protocol. F1000Res. Aug 13, 2021;10:802. [FREE Full text] [CrossRef] [Medline]
  123. Hu X, Rousseau R, Chen J. On the definition of forward and backward citation generations. J Informetr. Jan 2011;5(1):27-36. [FREE Full text] [CrossRef]
  124. Ouzzani M, Hammady H, Fedorowicz Z, Elmagarmid A. Rayyan-a web and mobile app for systematic reviews. Syst Rev. Dec 05, 2016;5(1):210. [FREE Full text] [CrossRef] [Medline]
  125. Levac D, Colquhoun H, O'Brien KK. Scoping studies: advancing the methodology. Implement Sci. Sep 20, 2010;5:69. [FREE Full text] [CrossRef] [Medline]
  126. Kuckartz U, Rädiker S. Coding text and PDF files. In: Kuckartz U, Rädiker S, editors. Analyzing Qualitative Data with MAXQDA: Text, Audio, and Video. Cham, Switzerland. Springer; 2019;65-81.
  127. DeLone WH, McLean ER. The DeLone and McLean model of information systems success: a ten-year update. J Manag Inf Syst. 2003;19(4):9-30. [FREE Full text] [CrossRef]
  128. Hevner A, Chatterjee S. Design science research: looking to the future. In: Hevner A, Chatterjee S, editors. Design Research in Information Systems: Theory and Practice. New York, NY, USA. Springer; 2010;261-268.
  129. Zhang J, Walji MF. TURF: toward a unified framework of EHR usability. J Biomed Inform. Dec 2011;44(6):1056-1067. [FREE Full text] [CrossRef] [Medline]
  130. Taylor MJ, McNicholas C, Nicolay C, Darzi A, Bell D, Reed JE. Systematic review of the application of the plan-do-study-act method to improve quality in healthcare. BMJ Qual Saf. Apr 2014;23(4):290-298. [FREE Full text] [CrossRef] [Medline]
  131. Celano G, Costa A, Fichera S, Tringali G. Linking six sigma to simulation: a new roadmap to improve the quality of patient care. Int J Health Care Qual Assur. 2012;25(4):254-273. [CrossRef] [Medline]
  132. Randall D, Harper R, Rouncefield M. Fieldwork for design. Comput Support Coop Work. 2007:89-131. [CrossRef]
  133. Srivastava A, Thomson SB. Framework analysis: a qualitative methodology for applied policy research. J Admin Gov. 2009;4(2):72-79. [FREE Full text]
  134. Ammenwerth E, de Keizer N. An inventory of evaluation studies of information technology in health care. Methods Inf Med. 2005;44(01):44-56. [FREE Full text] [CrossRef]
  135. Venkatesh V, Morris MG, Davis GB, Davis FD. User acceptance of information technology: toward a unified view. MIS Quarterly. Sep 2003;27(3):425-478. [FREE Full text] [CrossRef]
  136. Flanagan JC. The critical incident technique. Psychol Bull. Jul 1954;51(4):327-358. [CrossRef] [Medline]
  137. Greenhalgh T, Robert G, Macfarlane F, Bate P, Kyriakidou O. Diffusion of innovations in service organizations: systematic review and recommendations. Milbank Q. 2004;82(4):581-629. [FREE Full text] [CrossRef] [Medline]
  138. Holden RJ, Carayon P, Gurses AP, Hoonakker P, Hundt AS, Ozok AA, et al. SEIPS 2.0: a human factors framework for studying and improving the work of healthcare professionals and patients. Ergonomics. 2013;56(11):1669-1686. [FREE Full text] [CrossRef] [Medline]
  139. Lyell D, Magrabi F, Coiera E. The effect of cognitive load and task complexity on automation bias in electronic prescribing. Hum Factors. Nov 2018;60(7):1008-1021. [CrossRef] [Medline]
  140. Egeth H, Kahneman D. Attention and effort. Am J Psychol. Jun 1975;88(2):339. [FREE Full text] [CrossRef]
  141. Holden RJ, Karsh BT. The technology acceptance model: its past and its future in health care. J Biomed Inform. Feb 2010;43(1):159-172. [FREE Full text] [CrossRef] [Medline]
  142. Guérin F, Laville A, Daniellou F, Duraffourg J, Kerguelen A. Compreender o trabalho para transformá-lo: a prática da ergonomia. Sao Paulo, Brazil. Funcacao Vanzolini; 2001.
  143. Zheng K, Guo MH, Hanauer DA. Using the time and motion method to study clinical work processes and workflow: methodological inconsistencies and a call for standardized research. J Am Med Inform Assoc. Sep 2011;18(5):704-710. [FREE Full text] [CrossRef] [Medline]
  144. 10 usability heuristics for user interface design. Nielsen Norman Group. 1994. URL: https://www.nngroup.com/articles/ [accessed 2022-09-15]
  145. Boehm BW. A spiral model of software development and enhancement. Computer. May 1988;21(5):61-72. [FREE Full text] [CrossRef]
  146. Brown SH, Lincoln MJ, Groen PJ, Kolodner RM. VistA--U.S. department of veterans affairs national-scale HIS. Int J Med Inform. Mar 2003;69(2-3):135-156. [CrossRef] [Medline]
  147. Topaz M. The Hitchhiker’s guide to nursing informatics theory: using the data-knowledge-information-wisdom framework to guide informatics research. Online J Nurs Inform. 2013;17(3):2852-2857. [FREE Full text]
  148. Bush R, Pérez A, Baum T, Etland C, Connelly CD. A systematic review of the use of the electronic health record for patient identification, communication, and clinical support in palliative care. JAMIA Open. Oct 01, 2018;1(2):294-303. [FREE Full text] [CrossRef] [Medline]
  149. Kernebeck S, Busse TS, Ehlers JP, Vollmar HC. Adhärenz digitaler interventionen im gesundheitswesen: definitionen, methoden und offene fragen. Bundesgesundheitsblatt Gesundheitsforschung Gesundheitsschutz. Oct 2021;64(10):1278-1284. [FREE Full text] [CrossRef] [Medline]
  150. Jimma BL, Enyew DB. Barriers to the acceptance of electronic medical records from the perspective of physicians and nurses: a scoping review. Inform Med Unlocked. 2022;31:100991. [FREE Full text] [CrossRef]
  151. Saad-Sulonen J, Eriksson E, Halskov K, Karasti H, Vines J. Unfolding participation over time: temporal lenses in participatory design. CoDesign. Jan 31, 2018;14(1):4-16. [FREE Full text] [CrossRef]
  152. Khawaja MA, Chen F, Marcus N. Measuring cognitive load using linguistic features: implications for usability evaluation and adaptive interaction design. Int J Hum Comput Interact. Apr 2014;30(5):343-368. [FREE Full text] [CrossRef]
  153. Brandt E, Binder T, Sanders E. Tools and techniques: ways to engage telling, making and enacting. In: Simonsen J, Robertson T, editors. Routledge International Handbook of Participatory Design. New York, NY, USA. Routledge International; 2012;145-181.
  154. Kernebeck S, Scheibe M, Sinha M, Fischer F, Knapp A, Timpel P, et al. Development, evaluation and implementation of digital health interventions (Part 1) - discussion paper of the digital health working group of the German network for health services research (DNVF). Gesundheitswesen. Jan 2023;85(1):58-64. [CrossRef] [Medline]
  155. Kernebeck S, Scheibe M, Sinha M, Fischer F, Knapp A, Timpel P, et al. Developing, evaluating and implementing digital health interventions part II - discussion paper of the digital health working group of the German network for health services research (DNVF). Gesundheitswesen. Jan 2023;85(1):65-70. [CrossRef] [Medline]
  156. Moser A, Korstjens I. Series: practical guidance to qualitative research. Part 5: co-creative qualitative approaches for emerging themes in primary care research: experience-based co-design, user-centred design and community-based participatory research. Eur J Gen Pract. Dec 2022;28(1):1-12. [FREE Full text] [CrossRef] [Medline]
  157. Jones M, Palanque P, Schmidt A, Grossman T, Sas C, Whittaker S, et al. Generating implications for design through design research. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. Presented at: CHI '14; April 26-May 1, 2014, 2014;1971-1980; Toronto, Canada. URL: https://dl.acm.org/doi/10.1145/2556288.2557357 [CrossRef]
  158. Kernebeck S, Kramer U, Redaelli M, Vollmar HC. Bewerten, aber wie? – Kriterien für die Evaluation und das Reporting von Studien zu Gesundheits- und Medizin-Apps. Monitor Versorgungsforschung. Mar 30, 2021;14(02/2021):69-75. [FREE Full text] [CrossRef]


EHR: electronic health record
ICU: intensive care unit
ISO: International Organization for Standardization
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses


Edited by C Jacob; submitted 09.01.23; peer-reviewed by E Anderson, M Alkureishi; comments to author 25.02.23; revised version received 15.04.23; accepted 17.05.23; published 10.07.23.

Copyright

©Theresa Sophie Busse, Chantal Jux, Johannes Laser, Peter Rasche, Horst Christian Vollmar, Jan P Ehlers, Sven Kernebeck. Originally published in JMIR Human Factors (https://humanfactors.jmir.org), 10.07.2023.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Human Factors, is properly cited. The complete bibliographic information, a link to the original publication on https://humanfactors.jmir.org, as well as this copyright and license information must be included.