The extracted data included information about the study (eg, authors, year of publication, title, and country), population (eg, number of participants), concepts (usability methods, usability attributes, and usability phase), and context (educational setting). The final data extraction sheet can be found in Multimedia Appendix 3 [ 24 - 111 ]. One review author extracted the data from the included studies using Microsoft Excel software [ 21 ], which was checked by another researcher.
Descriptions of usability attributes have not been standardized, making categorization challenging. Therefore, a review author used deductive analysis to interpret the usability attributes reported in the included studies. This interpretation was based on a review of usability attributes as defined in previous literature. These definitions were assessed on the basis of the results of the included studies. This analysis was reviewed and discussed by another author. Disagreements were resolved through a consensus-based discussion.
Frequencies and percentages were used to present nominal data, together with tables and graphical illustrations. For instance, a figure showing the study selection process, an illustration of the frequency of inquiry-based usability evaluation and data collection methods, and an overview of the distribution of identified usability attributes were provided.
Database searches yielded 34,369 records, and 2796 records were identified using other methods. After removing duplicates, 28,702 records remained. A total of 626 reports were examined in full text. In all, 88 articles were included in the scoping review [ 24 - 111 ] ( Figure 1 ). A total of 8 articles comprised results from several studies in the same article, presented as study A, study B, or study C in Multimedia Appendix 3 . Therefore, a total of 98 studies were reported in the 88 articles included.
PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flowchart of study selection process.
The included studies comprised a total sample population of 7790, with participant numbers ranging from 5 to 736 participants per study. Most of the studies included medical students (34/88, 39%) or nursing students (25/88, 28%). Other participants included students from the following disciplines: pharmacy (9/88, 10%), dentistry (5/88, 6%), physiotherapy (5/88, 6%), health sciences (3/88, 3%), and psychology (2/88, 2%). Further information is provided in Multimedia Appendix 3 . There were 22 publishing countries, with most studies being from the United States (22/88, 25%), Spain (9/88, 10%), the United Kingdom (8/88, 9%), Canada (7/88, 8%), and Brazil (7/88, 8%), with an increasing number of publications from 2014. Table 2 provides an overview and characteristics of the included articles.
Characteristics of included articles.
Study number | Study | Population (N) | Research design: data collection method | Usability attributes |
1 | Aebersold et al [ ], 2018, United States | Nursing (N=69) | Mixed methods: questionnaire; task and knowledge performance | Ease of use; learning performance; satisfaction; usefulness |
2 | Akl et al [ ], 2008, United States | Resident (N=30) | Qualitative methods: focus groups; written qualitative reflections | Satisfaction |
3 | Al-Rawi et al [ ], 2015, United States | Dentist (N=61) | Posttest 1-group design: questionnaire | Ease of use; frequency of use; satisfaction; usefulness |
4 | Albrecht et al [ ], 2013, Germany | Medicine (N=6) | Posttest 1-group design: questionnaire | Satisfaction |
5 | Alencar Neto et al [ ], 2020, Brazil | Medicine (N=132) | Posttest 1-group design: questionnaire | Ease of use; learnability; satisfaction; usefulness |
6 | Alepis and Virvou [ ], 2010, Greece | Medicine (N=110) | Mixed methods: questionnaire; interviews | Ease of use; usefulness; user-friendliness |
7 | Ameri et al [ ], 2020, Iran | Pharmacy (N=241) | Posttest 1-group design: questionnaire | Context of use; efficiency; usefulness |
8 | Balajelini and Ghezeljeh [ ], 2018, Iran | Nursing (N=41) | Posttest 1-group design: questionnaire | Ease of use; frequency of use; navigation; satisfaction; simplicity; usefulness |
9 | Barnes et al [ ], 2015, United Kingdom | Medicine (N=42) | Randomized controlled trial: questionnaire; task and knowledge performance | Ease of use; effectiveness; learning performance; satisfaction |
10 | Busanello et al [ ], 2015, Brazil | Dentist (N=62) | Pre-post test, nonrandomized control group design: questionnaire | Learnability; learning performance; satisfaction |
11 | Cabero-Almenara and Roig-Vila [ ], 2019, Spain | Medicine (N=50) | Pre-post test, 1-group design: questionnaire | Learning performance; satisfaction |
12 | Choi et al [ ], 2015, South Korea | Nursing (N=5) | Think-aloud methods: interviews; data from app | Context of use; ease of use; learnability; satisfaction; usefulness |
13 | Choi et al [ ], 2018, South Korea | Nursing (N=75) | Pre-post test, nonrandomized control group design: questionnaire | Ease of use; learning performance; satisfaction; usefulness |
14 | Choo et al [ ], 2019, Singapore | Psychology (N=8) | Mixed methods: questionnaire ; written qualitative reflections | Ease of use; learning performance; satisfaction; usefulness; user-friendliness |
15 | Chreiman et al [ ], 2017, United States | Medicine (N=30) | Posttest 1-group design: questionnaire; data from app | Context of use; ease of use; frequency of use; usefulness |
16 | Colucci et al [ ], 2015, United States | Medicine (N=115) | Posttest 1-group design: questionnaire | Effectiveness; efficiency; satisfaction; usefulness |
17 | Davids et al [ ], 2014, South Africa | Residents (N=82) | Randomized controlled trial: questionnaire ; data from app | Effectiveness; efficiency; learnability; navigation; satisfaction; user-friendliness |
18A | Demmans et al [ ], 2018, Canada | Nursing (N=60) | Pre-post test, nonrandomized control group design: questionnaire; observations | Ease of use; effectiveness; learnability; learning performance; navigation; satisfaction |
18B | Demmans et al [ ], 2018, Canada | Nursing (N=85) | Pre-post test, nonrandomized control group design: questionnaire; observations | Ease of use; effectiveness; learnability; learning performance; navigation; satisfaction |
19 | Devraj et al [ ], 2021, United States | Pharmacy (N=89) | Posttest 1-group design: questionnaire; data from app | Ease of use; errors; frequency of use; learning performance; navigation; operational usability; satisfaction; usefulness |
20 | Díaz-Fernández et al [ ], 2016, Spain | Physiotherapy (N=110) | Posttest 1-group design: questionnaire | Comprehensibility; ease of use; usefulness |
21 | Docking et al [ ], 2018, United Kingdom | Paramedic (N=24) | Think-aloud methods: focus groups | Context of use; learnability; satisfaction; usefulness |
22 | Dodson and Baker [ ], 2020, United States | Nursing (N=23) | Qualitative methods: focus groups | Ease of use; operational usability; satisfaction; usefulness; user-friendliness |
23 | Duarte Filho et al [ ], 2014, Brazil | Medicine (N=10) | Posttest nonrandomized control group design: questionnaire | Ease of use; efficiency; satisfaction; usefulness |
24 | Duggan et al [ ], 2020, Canada | Medicine (N=80) | Posttest 1-group design: questionnaire; data from app | Ease of use; frequency of use; satisfaction; usefulness |
25 | Fernandez-Lao et al [ ], 2016, Spain | Physiotherapy (N=49) | Randomized controlled trial: questionnaire ; task and knowledge performance | Learning performance; satisfaction |
26 | Fralick et al [ ], 2017, Canada | Medicine (N=62) | Pre-post test, nonrandomized control group design: questionnaire | Ease of use; frequency of use; learning performance; usefulness |
27 | Ghafari et al [ ], 2020, Iran | Nursing (N=8) | Posttest 1-group design: questionnaire | Ease of use; operational usability; satisfaction; usefulness |
28 | Goldberg et al [ ], 2014, United States | Medicine (N=18) | Posttest 1-group design: questionnaire; task and knowledge performance | Ease of use; effectiveness |
29 | Gutiérrez-Puertas et al [ ], 2021, Spain | Nursing (N=184) | Randomized controlled trial: questionnaire; task and knowledge performance | Learning performance; satisfaction |
30 | Herbert et al [ ], 2021, United States | Nursing (N=33) | Randomized controlled trial: questionnaire; task and knowledge performance | Ease of use; learning performance; navigation; operational usability; usefulness |
31 | Hsu et al [ ], 2019, Taiwan | Nursing (N=16) | Qualitative methods: interviews | Context of use; operational usability; satisfaction; usefulness |
32 | Huang et al [ ], 2010, Taiwan | Not clear (N=28) | Posttest 1-group design: questionnaire | Ease of use; satisfaction, usefulness |
33 | Hughes and Kearney [ ], 2017, United States | Occupational therapy (N=19) | Qualitative methods: focus groups | Efficiency; satisfaction |
34 | Ismail et al [ ], 2018, Malaysia | Health science (N=124) | Pre-post test, 1-group design: questionnaire | Ease of use; learning performance; satisfaction; user-friendliness |
35 | Johnson et al [ ], 2021, Norway | Occupational therapy, physiotherapy, and social education (N=15) | Qualitative methods: focus groups | Context of use; ease of use; operational usability |
36A | Kang Suh [ ], 2018, South Korea | Nursing (N=92) | Pre-post test, nonrandomized control group design: questionnaire; data from app | Effectiveness; frequency of use; learning performance; satisfaction |
36B | Kang Suh [ ], 2018, South Korea | Nursing (N=49) | Qualitative methods: focus groups | Effectiveness; frequency of use; learning performance; satisfaction |
37 | Keegan et al [ ], 2016, United States | Nursing (N=116) | Posttest nonrandomized control group design: questionnaire; task and knowledge performance | Learning performance; satisfaction; usefulness |
38 | Kim-Berman et al [ ], 2019, United States | Dentist (N=93) | Posttest 1-group design: questionnaire; task and knowledge performance | Context of use; ease of use; effectiveness; usefulness |
39 | Kojima et al [ ], 2011, Japan | Physiotherapy and occupational therapy (N=41) | Pre-post test, 1-group design: questionnaire | Ease of use; learning performance; satisfaction; usefulness |
40 | Koulias et al [ ], 2012, Australia | Medicine (N=171) | Posttest 1-group design: questionnaire | Ease of use; operational usability; satisfaction |
41 | Kow et al [ ], 2016, Singapore | Medicine (N=221) | Pre-post test, 1-group design: questionnaire | Learning performance; satisfaction |
42 | Kurniawan and Witjaksono [ ], 2018, Indonesia | Medicine (N=30) | Posttest 1-group design: questionnaire | Satisfaction; usefulness |
43A | Lefroy et al [ ], 2017, United Kingdom | Medicine (N=21) | Qualitative methods: focus groups; data from app | Context of use; frequency of use; satisfaction |
43B | Lefroy et al [ ], 2017, United Kingdom | Medicine (N=405) | Quantitative methods: data from app | Context of use; frequency of use; satisfaction |
44 | Li et al [ ], 2019, Taiwan | Health care (N=70) | Pre-post test, nonrandomized control group design: questionnaire | Ease of use; usefulness |
45 | Lin and Lin [ ], 2016, Taiwan | Nursing (N=36) | Pre-post test, nonrandomized control group design: questionnaire | Cognitive load; ease of use; learnability; learning performance; usefulness |
46 | Lone et al [ ], 2019, Ireland | Dentist (N=59) | Randomized controlled trial: questionnaire; task and knowledge performance | Ease of use; learnability; learning performance; operational usability; satisfaction |
47A | Long et al [ ], 2016, United States | Nursing (N=158) | Pre-post test, 1-group design: questionnaire; data from app | Ease of use; efficiency; learnability; learning performance; satisfaction |
47B | Long et al [ ], 2016, United States | Health science (N=159) | Randomized controlled trial: questionnaire; data from app | Ease of use; efficiency; learnability; learning performance; satisfaction |
48 | Longmuir [ ], 2014, United States | Medicine (N=56) | Posttest 1-group design: questionnaire; data from app | Efficiency; learnability; operational usability; satisfaction |
49 | López et al [ ], 2016, Spain | Medicine (N=67) | Posttest 1-group design: questionnaire | Context of use; ease of use; errors; satisfaction; usefulness |
50 | Lozano-Lozano et al [ ], 2020, Spain | Physiotherapy (N=110) | Randomized controlled trial: questionnaire; task and knowledge performance | Learning performance; satisfaction; usefulness |
51 | Lucas et al [ ], 2019, Australia | Pharmacy (N=39) | Pre-post test, 1-group design: questionnaire; task and knowledge performance | Satisfaction; usefulness |
52 | Mathew et al [ ], 2014, Canada | Medicine (N=5) | Think-aloud methods: questionnaire ; interviews; task and knowledge performance | Learnability; satisfaction |
53 | McClure [ ], 2019, United States | Nursing (N=16) | Posttest 1-group design: questionnaire | Learnability; satisfaction; usefulness |
54 | McDonald et al [ ], 2018, Canada | Medicine (N=20) | Pre-post test, 1-group design: questionnaire; data from app | Effectiveness; satisfaction |
55 | McLean et al [ ], 2014, Australia | Medicine (N=58) | Mixed methods: questionnaire; focus groups; interviews | Satisfaction |
56 | McMullan [ ], 2018, United Kingdom | Health science (N=60) | Pre-post test, 1-group design: questionnaire | Learning performance; navigation; satisfaction; usefulness; user-friendliness |
57 | Mendez-Lopez et al [ ], 2021, Spain | Psychology (N=67) | Pre-post test, 1-group design: questionnaire; task and knowledge performance | Cognitive load; ease of use; learning performance; satisfaction; usefulness |
58 | Meruvia-Pastor et al [ ], 2016, Canada | Nursing (N=10) | Pre-post test, 1-group design: questionnaire; task and knowledge performance | Ease of use; learning performance; satisfaction; usefulness |
59 | Mettiäinen [ ], 2015, Finland | Nursing (N=121) | Mixed methods: questionnaire; focus groups | Ease of use; usefulness |
60 | Milner et al [ ], 2020, United States | Medicine and nursing (N=66) | Posttest 1-group design: questionnaire | Satisfaction; usefulness |
61 | Mladenovic et al [ ], 2021, Serbia | Dentist (N=56) | Posttest 1-group design: questionnaire | Context of use; ease of use; satisfaction; usefulness |
62 | Morris and Maynard [ ], 2010, United Kingdom | Physiotherapy and nursing (N=19) | Pre-post test, 1-group design: questionnaire | Context of use; ease of use; navigation; operational usability; usefulness |
63A | Nabhani et al [ ], 2020, United Kingdom | Pharmacy (N=56) | Posttest 1-group design: questionnaire | Ease of use; learnability; learning performance; satisfaction; usefulness |
63B | Nabhani et al [ ], 2020, United Kingdom | Pharmacy (N=152) | Posttest 1-group design: questionnaire | Ease of use; learnability; learning performance; satisfaction; usefulness |
63C | Nabhani et al [ ], 2020, United Kingdom | Pharmacy (N=33) | Posttest 1-group design: task and knowledge performance | Ease of use; learnability; learning performance; satisfaction; usefulness |
64A | Noguera et al [ ], 2013, Spain | Physiotherapy (N=84) | Posttest 1-group design: questionnaire | Learning performance; satisfaction; usefulness |
64B | Noguera et al [ ], 2013, Spain | Physiotherapy (N=76) | Randomized controlled trial: questionnaire | Learning performance; satisfaction; usefulness |
65 | O’Connell et al [ ], 2016, Ireland | Medicine, nursing, and pharmacy (N=89) | Randomized controlled trial: questionnaire | Ease of use; learning performance; operational usability; satisfaction; simplicity |
66 | Oliveira et al [ ], 2019, Brazil | Medicine (N=110) | Randomized controlled trial: questionnaire; task and knowledge performance | Frequency of use; learning performance; satisfaction |
67 | Orjuela et al [ ], 2015, Colombia | Medicine (N=22) | Posttest 1-group design: questionnaire | Ease of use; satisfaction |
68 | Page et al [ ], 2016, United States | Medicine (N=356) | Mixed methods: questionnaire; interviews | Context of use; efficiency; satisfaction |
69 | Paradis et al [ ], 2018, Canada | Medicine and nursing (N=108) | Posttest 1-group design: questionnaire | Ease of use; satisfaction; usefulness |
70 | Pereira et al [ ], 2017, Brazil | Medicine (N=20) | Posttest 1-group design: questionnaire | Ease of use; learnability; satisfaction; usefulness |
71 | Pereira et al [ ], 2019, Brazil | Nursing (N=60) | Posttest 1-group design: questionnaire | Ease of use; operational usability; satisfaction |
72A | Pinto et al [ ], 2008, Brazil | Biomedical informatics (N=5) | Qualitative methods: observations; task and knowledge performance | Efficiency; errors; learnability; learning performance; operational usability; satisfaction |
72B | Pinto et al [ ], 2008, Brazil | Medicine (N=not clear) | Posttest nonrandomized control group design: questionnaire | Efficiency; errors; learnability; learning performance; operational usability; satisfaction |
73 | Quattromani et al [ ], 2018, United States | Nursing (N=181) | Randomized controlled trial: questionnaire | Learnability; learning performance; satisfaction; usefulness |
74 | Robertson and Fowler [ ], 2017, United States | Medicine (N=18) | Qualitative methods: focus groups | Satisfaction |
75A | Romero et al [ ], 2021, Germany | Medicine (N=22) | Think-aloud methods: questionnaire; interviews; task and knowledge performance | Effectiveness; efficiency; errors; navigation; satisfaction |
75B | Romero et al [ ], 2021, Germany | Medicine (N=22) | Posttest 1-group design: questionnaire | Learnability; satisfaction |
75C | Romero et al [ ], 2021, Germany | Medicine (N=736) | Posttest 1-group design: questionnaire | Frequency of use; satisfaction |
76 | Salem et al [ ], 2020, Australia | Pharmacy (N=33) | Posttest 1-group design: questionnaire | Operational usability; satisfaction; usefulness |
77 | San Martín-Rodríguezet al [ ], 2020, Spain | Nursing (N=77) | Posttest 1-group design: questionnaire; task and knowledge performance | Learning performance; operational usability; satisfaction |
78 | Schnepp and Rogers [ ], 2017, United States | Not clear (N=72) | Think-aloud methods: questionnaire ; interviews; task and knowledge performance | Learnability; satisfaction |
79 | Smith et al [ ], 2016, United Kingdom | Medicine and nursing (N=74) | Mixed methods: questionnaire; focus groups | Navigation; operational usability; satisfaction; user-friendliness |
80 | Strandell-Laine et al [ ], 2019, Finland | Nursing (N=52) | Mixed methods: questionnaire ; written qualitative responses | Learnability; operational usability; satisfaction |
81 | Strayer et al [ ], 2010, United States | Medicine (N=122) | Mixed methods: questionnaire; focus groups | Context of use; learnability; learning performance; satisfaction; usefulness |
82 | Taylor et al [ ], 2010, United Kingdom | A total of 8 different health care educations (N=79) | Qualitative methods: focus groups; written qualitative reflections | Context of use; learnability |
83 | Toh et al [ ], 2014, Singapore | Pharmacy (N=31) | Posttest 1-group design: questionnaire | Ease of use; learnability; navigation; usefulness |
84 | Tsopra et al [ ], 2020, France | Medicine (N=57) | Mixed methods: questionnaire; focus groups | Ease of use; operational usability; satisfaction; usefulness |
85 | Wu [ ], 2014, Taiwan | Nursing (N=36) | Mixed methods: questionnaire; interviews | Cognitive load; effectiveness; satisfaction; usefulness |
86 | Wyatt et al [ ], 2012, United States | Nursing (N=12) | Qualitative methods: focus groups | Ease of use; efficiency; errors; learnability; memorability; navigation; satisfaction |
87 | Yap [ ], 2017, Singapore | Pharmacy (N=123) | Posttest 1-group design: questionnaire | Comprehensibility; learning performance; memorability; navigation; satisfaction; usefulness |
88 | Zhang et al [ ], 2015, Singapore | Medicine (N=185) | Mixed methods: questionnaire; focus groups | Usefulness |
a Performances measured, comparing paper and app results, quiz results, and exam results.
b Reported use of validated questionnaires.
The usability evaluation methods found were either inquiry-based or based on user testing. The following inquiry methods were used: 1-group design (46/98, 47%), control group design (12/98, 12%), randomized controlled trials (12/98, 12%), mixed methods (12/98, 12%), and qualitative methods (11/98, 11%). Several studies that applied inquiry-based methods used more than one data collection method, with questionnaires being used most often (80/98, 82%), followed by task and knowledge performance testing (17/98, 17%), focus groups (15/98, 15%), collection of user data from the app (10/98, 10%), interviews (5/98, 5%), written qualitative reflections (4/98, 4%), and observations (3/98, 3%). Additional information can be found in the data extraction sheet ( Multimedia Appendix 3 ). Figure 2 illustrates the frequency of the inquiry-based usability evaluation methods and data collection methods.
Inquiry usability evaluation methods and data collection methods.
The only user testing methods found were think-aloud methods (5/98, 5%), and 4 (80%) of these studies applied more than one data collection method. The data collection methods used included interviews (4/98, 4%), questionnaires (3/98, 3%), task and knowledge performance (3/98, 3%), focus groups (1/98, 1%), and collection of user data from the app (1/98, 1%).
A total of 19 studies used a psychometrically tested usability questionnaire, including the SUS, Technology Acceptance Model, Technology Satisfaction Questionnaire, and Technology Readiness Index. SUS [ 112 ] was used in most (9/98, 9%) of the studies.
Field testing was the most frequent type of usability experiment, accounting for 72% (71/98) of usability experiments. A total of 22 (22%) studies performed laboratory testing, and 5 (5%) studies did not indicate the type of experiment performed. Multimedia Appendix 3 provides an overview of the type of experiment conducted in each study. The usability testing of the mobile apps took place in a classroom setting (41/98, 42%), in clinical placement (29/98, 30%), during simulation training (14/98, 14%), other (7/98, 7%), or the setting was not specified (5/98, 5%).
A total of 17 usability attributes have been identified among the included studies. The most frequently identified attributes were satisfaction, usefulness, ease of use, learning performance, and learnability. The least frequent were errors, cognitive load, comprehensibility, memorability, and simplicity. Table 3 provides an overview of the usability attributes identified in the included studies.
Distribution of usability attributes (n=17) and affiliated reports (N=88).
Usability attribute | Distribution, n (%) | Reports (references) |
Satisfaction | 74 (84) | [ - , - , - , - , , , - , , , - , - , , , - , - ] |
Usefulness | 51 (58) | [ , , - , - , - , , , - , - , , , , - , , - , , , , , , - , , ] |
Ease of use | 45 (51) | [ , , , , , , - , - , - , - , , , , , - , - , , - , - , , , - , , , ] |
Learning performance | 33 (38) | [ , - , , , , , , , , , , , , , , - , , - , - , , , , , ] |
Learnability | 23 (26) | [ , , , , , , - , , , , , , , , , - , ] |
Operational usability | 19 (22) | [ , , , , , , , , , , , , , , - , , ] |
Context of use | 14 (16) | [ , , , , , , , , , , , , , ] |
Navigation | 12 (14) | [ , - , , , , , , , , ] |
Efficiency | 11 (13) | [ , , , , , , , , , , ] |
Effectiveness | 10 (11) | [ , - , , , , , , ] |
Frequency of use | 10 (11) | [ , , , , , , , , , ] |
User-friendliness | 7 (8) | [ , , , , , , ] |
Errors | 5 (6) | [ , , , , ] |
Cognitive load | 3 (3) | [ , , ] |
Comprehensibility | 2 (2) | [ , ] |
Memorability | 2 (2) | [ , ] |
Simplicity | 2 (2) | [ , ] |
This scoping review sought to identify the usability methods and attributes reported in usability studies of mobile apps for health care education. A total of 88 articles, with a total of 98 studies reported in these 88 articles, were included in this review. Our findings indicate a steady increase in publications from 2014, with studies being published in 22 different countries. Field testing was used more frequently than laboratory testing. Furthermore, the usability evaluation methods applied were either inquiry-based or based on user testing. Most of the inquiry-based methods were experiments that used questionnaires as a data collection method, and all of the studies with user testing methods applied think-aloud methods. Satisfaction, usefulness, ease of use, learning performance, and learnability were the most frequently identified usability attributes.
The studies included in this scoping review mainly applied inquiry-based methods, primarily the collection of self-reported data through questionnaires. This is congruent with the results of Weichbroth [ 10 ], in which controlled observations and surveys were the most frequently applied methods. Asking users to respond to a usability questionnaire may provide relevant and valuable information. Among the 83 studies that used questionnaires in our review, only 19 (23%) used a psychometrically tested usability questionnaire; of these, the SUS questionnaire [ 112 ] was used most frequently. In line with the review on usability questionnaires [ 12 ], we recommend using a psychometrically tested usability questionnaire to support the advancement of usability science. As questionnaires address only certain usability attributes, mainly learnability, efficiency, and satisfaction [ 12 ], it would be helpful to also include additional methods, such as interviews or mixed methods, and to incorporate additional open-ended questions when using questionnaires.
Furthermore, the application of usability evaluation methods other than inquiry methods, such as user testing methods and inspection methods [ 10 ], could be beneficial and lead to more objective measures of app usability. Among other things, subjective data are collected via self-reported questionnaires, and objective data are collected based on task completion rates [ 40 ]. For example, in one of the included studies, the participants reported that the usability of the app was satisfactory by subjective measures, but the participants did not use the app [ 75 ]. Another study reported a lack of coherence between subjective and objective data; thus, these results indicate the importance of not relying solely on subjective measures of usability [ 40 ]. Therefore, it is suggested that various usability evaluation methods, including subjective and objective usability measures, are used in future usability studies.
Our review found that most of the included studies in health care education (71/98, 72%) performed field testing, whereas previous literature suggests that usability experiments in other fields are more often conducted in a laboratory [ 1 , 113 ]. For instance, Kumar and Mohite [ 1 ] found that 73% of the studies included in their review of mobile learning apps used laboratory testing. Mobile apps in health care education have been developed to support students’ learning, on-campus and during clinical placement, in various settings and on the move. Accordingly, it is especially important to test how the apps are perceived in specific environments [ 5 ]; hence, field testing is required. However, many usability issues can be discovered in a laboratory. Particularly in the early phases of app development, testing an app with several participants in a laboratory may make it more feasible to test and improve the app [ 8 ]. Usability testing in a laboratory can provide rapid feedback on usability issues, which can then be addressed before testing the app in a real-world environment. Therefore, it may be beneficial to conduct small-scale laboratory testing before field testing.
Previous systematic reviews of mobile apps in general identified satisfaction, efficiency, and effectiveness as the most common usability attributes [ 5 , 10 ]. In this review, efficiency and effectiveness were explored to a limited extent, whereas satisfaction, usefulness, and ease of use were the most frequently identified usability attributes. Our results coincide with those from a previous review on the usability of mobile learning apps [ 1 ], possibly because satisfaction, usefulness, and ease of use are usability attributes of particular importance when examining mobile learning apps.
Learning performance was assessed frequently in the included studies. For ensuring that apps are valuable in a given learning context, it is relevant to test additional usability attributes such as cognitive load [ 9 ]. However, few studies included in our review examined cognitive load [ 68 , 80 , 108 ]. Mobile apps are often used in an environment with multiple distractions, which may contribute to an increased cognitive load [ 5 ], affecting the learning performance. Testing both learning performance and app users’ cognitive load may improve the understanding of the app’s usability.
We found that several of the included studies did not use terminology from usability literature to describe which usability attributes they were testing. For instance, studies that tested satisfaction often used words such as “likes and dislikes” and “recommend use to others” and did not specify that they tested the usability attribute satisfaction. Specifying which usability attributes are investigated will be important when performing a usability study of mobile apps, as this will influence transparency and enable comparison between different studies. In addition, evaluating a wider range of usability attributes may enable researchers to expand their perspective regarding the app’s usability problems and ensure quicker improvement of the app. Defining and presenting different usability attributes in a reporting guideline can assist in deciding on and reporting relevant usability attributes. As such, a reporting guideline would be beneficial for researchers planning and conducting usability studies, a point that is also supported by the systematic review conducted by Kumar and Mohite [ 1 ].
Combining different usability evaluation methods that incorporate both subjective and objective usability measures can add various and important perspectives when developing apps. In future studies, it would be advantageous to use psychometrically tested usability questionnaires to support the advancement of the usability science. In addition, developers of mobile apps should determine which usability attributes are relevant before conducting usability studies (eg, by registering a protocol). Incorporating these perspectives into the development of a reporting guideline would be beneficial to future usability studies.
First, the search strategy was designed in collaboration with a research librarian and peer reviewed by another research librarian and included 10 databases and other sources. This broad search strategy resulted in a high number of references, which may be associated with a lower level of precision. To ensure the retrieval of all potentially pertinent articles, two of the authors independently screened titles and abstracts; studies deemed eligible by one of the authors were included for full-text screening.
Second, the full-text evaluation was challenging because the term usability has multiple meanings that do not always relate to usability testing. For instance, the term was used when testing students’ experience of a commercially developed app but not in connection with the app’s further development. In addition, many studies did not explicitly state that a mobile app was being investigated, which also created a challenge when deciding whether they satisfied the eligibility criteria. Nevertheless, reading the full-text articles independently by 2 reviewers and solving disagreements through consensus-based discussions ensured the inclusion of relevant articles.
This scoping review was performed to provide an overview of the usability methods used and the attributes identified in usability studies of mobile apps in health care education. Experimental designs were commonly used to evaluate usability and most studies used field testing. Questionnaires were frequently used for data collection, although few studies used psychometrically tested questionnaires. Usability attributes identified most often were satisfaction, usefulness, and ease of use. The results indicate that combining different usability evaluation methods, incorporating both subjective and objective usability measures, and specifying which usability attributes to test seem advantageous. The results can support the planning and conduct of future usability studies of the advancement of learning apps in health care education.
The research library at Western Norway University of Applied Sciences provided valuable assistance in developing and performing the search strategy for this scoping review. Gunhild Austrheim, a research librarian, provided substantial guidance in the planning and performance of the database searches. Marianne Nesbjørg Tvedt peer reviewed the search string. Malik Beglerovic also assisted with database searches. The authors would also like to thank Ane Kjellaug Brekke Gjerland for assessing the data extraction sheet.
PRISMA-ScR | Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews |
SUS | System Usability Scale |
Multimedia appendix 2, multimedia appendix 3.
Authors' Contributions: SGJ, LL, DC, and NRO proposed the idea for this review. SGJ, DC, and NRO contributed to the screening of titles and abstracts, and SGJ and TP decided on eligibility based on full-text examinations. SGJ extracted data from the included studies. SGJ, TP, LL, DC, and NRO contributed to the drafts of the manuscript, and all authors approved the final version for publication.
Conflicts of Interest: None declared.
IMAGES
COMMENTS
Usability evaluation methods: a literature review February 2012 International Journal of Engineering Science and Technology 4 (2) February 2012 4 (2) License CC BY 4.0 Authors: Ankita Madan
This paper presents a comprehensive study of different usability evaluation methods. The objective of this paper is to lay down the intensive and conceptual study of the usability concepts.
This chapter aims to identify, analyze, and classify the methodologies and methods described in the literature for the usability evaluation of systems and services based on information and communication technologies. The methodology used was a systematic review of the literature.
Conclusions: In summary, this paper provides a review of the usability evaluation methods employed in the assessment of eHealth HIV eHealth interventions. eHealth is a growing platform for delivery of HIV interventions and there is a need to critically evaluate the usability of these tools before deployment.
Abstract. This chapter aims to identify, analyze, and classify the methodologies and methods described in the literature for the usability evaluation of systems and services based on information ...
The usability evaluation process of mobile applications is carried out with a systematic literature review of 22 papers. The results show that 73% of the methods used are usability testing, 23% heuristic evaluations, and 4% are user satisfaction usability evaluations.
A Data Extraction Form (see Appendix A) was used to manage the review results, including (1) author (s) of the paper; (2) year of publication; (3) suggested usability evaluation methods, approaches, and models; (4) category of the mobile app; (5) usability attributes and features used for mobile app design and evaluation; and (6) usability ...
For included studies, we recorded usability evaluation methods or usability metrics as appropriate, and any measurement techniques applied to illustrate these. We classified and described all usability evaluation methods, usability metrics, and measurement techniques. Study quality was evaluated using a modified Downs and Black checklist.
The goal of this systematic mapping study was to examine the current use of usability evaluation methods in Web development. The principal findings of our study are the following: -. Usability evaluation methods have been constantly modified to better support the evaluation of Web artifacts.
Since usability is considered as a critical success factor for any software application, several evaluation methods have been developed. Nowadays, it is possible to find many proposals in the literature that address to evaluate usability issues. However, there is still discussion about what usability evaluation method is the most widely accepted by the scientific community. In this research, a ...
The exclusion criteria were as follows: (1) non-English studies, (2) focusing on only dashboard design or dashboard evaluation, (3) use of evaluation methods other than questionnaires to evaluate usability, and (4) lack of access to the full text of articles. 2.3. Study Selection, Article Evaluation, and Data Extraction.
Ideally usability testing should take place iteratively throughout the design of the resource, and there are several approaches for undertaking usability testing described in the wider literature. Within radiation oncology education, the extent to which usability testing occurs remains unclear.
Several types of usability evaluation methods (UEM) are used to assess software, and more extensive research is needed on the use of UEM in early design and development stages by manufacturers to achieve the goal of user-centered design. This article is a literature review of the most commonly applied UEM and related emerging trends.
Gathering users' design feedback as part of usability evaluation may be seen as controversial, and the current knowledge on users' design feedback is fragmented. To mitigate this, we have conducted a literature review. The review provides an overview of the benefits and limitations of users' design feedback in usability evaluations.
Therefore, a systematic literature review was conducted in order to identify usability evaluation guidelines for mobile educational games, which are concerning primary school students as users. This work is the first step toward making a set of usability guidelines for the evaluation of mobile educational games for Primary school students.
A literature review, outlined in the following section, was conducted as validation of the PACMAD model. This literature review examined which attributes of usability, as defined in the PACMAD usability model, were used during the evaluation of mobile applications presented in a range of papers published between 2008 and 2010. Previous work by Kjeldskov & Graham [ 3] has looked at the research ...
Usability evaluation is the evaluation of the product or system context of use, which is determined by the users, environment, tasks, and equipment. As the field of usability evaluation research has evolved, researchers have developed a variety of ways to apply the evaluation of usability methods. This systematic review aims to identify topics, trends, categories, methods and to answer ...
Through illustrative examples identified in the literature review, we demonstrate that usability testing is feasible and beneficial for educational resources varying in size and context. In doing so we hope to encourage radiation oncologists to incorporate usability testing into future educational resource design.
This review is a synthesis of research project about Information Ergonomics and embraces three dimensions, namely the methods, models and frameworks that have been applied to evaluate LMS. The study also includes the main usability criteria and heuristics used. The obtained results show a notorious change in the paradigms of usability, with ...
Literature survey is the dominant data collection technique for developing usability guidelines, while usability evaluation is the most common technique for validating newly developed guidelines. The findings concluded that there is a lack of standardization for developing, implementing, and evaluating usability guidelines.
In this article, we present a comprehensive review of the relevant literature. We analyze and compare publications of GIS usability evaluations, the methods that were used, and the findings reported. We thus produce a more detailed picture of GIS usability evaluations.
Conclusion We conducted a systematic review and expert-validation to identify rapidly deployable eHealth usability evaluation methods. The comprehensive and evidence-based prioritization of eHealth usability evaluation methods supports faster usability evaluations, and so contributes to the ease-of-use of emerging eHealth systems.
Person-based design and evaluation of MIA, a digital medical interview assistant for radiology ... its system architecture and the results from a comprehensive evaluation of the system including usability assessment. 2 Methods. ... which were derived from a narrative literature review and a patient survey, and supplemented by specifications ...
The study re-examined the space-social security role of complex features in communities during normal and epidemic periods and developed a three-level evaluation system using methods including literature crawling, high-frequency screening, and hierarchical analysis.
Background Post-stroke depression (PSD) is closely associated with poor stroke prognosis. However, there are some challenges in identifying and assessing PSD. This study aimed to identify scales for PSD diagnosis, assessment, and follow-up that are straightforward, accurate, efficient, and reproducible. Methods A systematic literature search was conducted in 7 electronic databases from January ...
Objective The aim of this scoping review is to identify usability methods and attributes in usability studies of mobile apps for health care education.