Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 02 February 2022
Sec. Quantitative Psychology and Measurement

Chinese Version of the mHealth App Usability Questionnaire: Cross-Cultural Adaptation and Validation

\r\nShuqing Zhao,Shuqing Zhao1,3Yingjuan Cao,,*Yingjuan Cao2,3,4*Heng Cao,Heng Cao3,5Kao Liu,Kao Liu2,4Xiaoyan Lv,Xiaoyan Lv2,4Jinxin ZhangJinxin Zhang3Yuxin LiYuxin Li3Patricia M. Davidson\r\nPatricia M. Davidson6
  • 1Department of Nursing, The Second Hospital, Cheeloo College of Medicine, Shandong University, Jinan, China
  • 2Department of Nursing, Qilu Hospital, Cheeloo College of Medicine, Shandong University, Jinan, China
  • 3School of Nursing and Rehabilitation, Shandong University, Jinan, China
  • 4Nursing Theory and Practice Innovation Research Center, Shandong University, Jinan, China
  • 5Department of Nursing, Shandong Provincial Hospital Affiliated to Shandong First Medical University, Jinan, China
  • 6The Vice-Chancellor’s Unit, University of Wollongong, Wollongong, NSW, Australia

Background: Mobile health (mHealth) apps have shown the advantages of improving medication compliance, saving time required for diagnosis and treatment, reducing medical expenses, etc. The World Health Organization (WHO) has recommended that mHealth apps should be evaluated prior to their implementation to ensure their accuracy in data analysis.

Objective: This study aimed to translate the patient version of the interactive mHealth app usability questionnaire (MAUQ) into Chinese, and to conduct cross-cultural adaptation and reliability and validity tests.

Methods: The Brislin’s translation model was used in this study. The cross-cultural adaptation was performed according to experts’ comments and the results of prediction test. The convenience sampling method was utilized to investigate 346 patients who used the “Good Doctor” (“Good Doctor” is the most popular mHealth app in China), and the reliability and validity of the questionnaire were evaluated as well.

Results: After translation and cross-cultural adaptation, there were a total of 21 items and 3 dimensions: usability and satisfaction (8 items), system information arrangement (6 items), and efficiency (7 items). The content validity index was determined to be 0.952, indicating that the 21 items used to evaluate the usability of the Chinese version of the MAUQ were well correlated. The Cronbach’s α coefficient of the total questionnaire was 0.912, which revealed that the questionnaire had a high internal consistency. The values of test-retest reliability and split-half reliability of the Chinese version of the MAUQ were 0.869 and 0.701, respectively, representing that the questionnaire had a good stability.

Conclusion: The translated questionnaire has good reliability and validity in the context of Chinese culture, and it could be used as a usability testing tool for the patient version of interactive mHealth apps.

Introduction

The Global Observatory for eHealth of the World Health Organization (WHO) defined mHealth as “medical and public health practice supported by mobile devices, such as mobile phones, patient monitoring devices, personal digital assistants, and other wireless devices.” A large number of investigations and studies on mHealth apps have shown that mHealth apps can improve medication compliance, save time required for diagnosis and treatment, and reduce medical expenses (Seto et al., 2012; Fairman et al., 2013; Parmanto et al., 2013; Pfammatter et al., 2016). With the aging of the population, demand for healthcare services has markedly increased, imposing a huge pressure on medical institutions (Association of American Medical Colleges [AAMC], 2021). According to a previous research, the Internet-based treatment could noticeably reduce the number of outpatient visits (National Health Service [NHS], 2019). It could also decrease the burden on hospital resources, medical staff’s workload, and medical expenses (Shah et al., 2021). Besides, mHealth apps have shown a potential to greatly improve healthcare systems.

China has a large population and relatively few medical resources. Faced with the fact that the development of mHealth apps can relieve the tension of medical institutions, the government has put forward the development of the Internet-based healthcare services. The coronavirus disease 2019 (Covid-19) pandemic has caused the urgent need to redesign public health systems from reactive to proactive and to develop innovations that may provide real-time information for making proactive decisions. Owing to the limited healthcare resources, several national healthcare providers are turning to digital healthcare solutions, such as virtual ward, remote patient monitoring, and telemedicine, to minimize the risk of Covid-19, in which mHealth apps have shown a noticeable efficiency (Mann et al., 2020; Schinkothe et al., 2020).

According to the published statistics, more than 325,000 mHealth apps have been universally developed in recent years (Maramba et al., 2019). However, nearly half of the million mHealth programs that have been developed are not extensively utilized. Because there was no effective usability evaluation of mHealth apps, clinicians, scholars, and patients are skeptical about the reliability of mHealth programs (Vera et al., 2019). These limitations have negatively influenced the applicability of mHealth apps. Hence, the WHO has recommended that mHealth apps should be evaluated prior to their actual use in order to ensure the accuracy in data analysis (World Health Organization [WHO], 2020). At the initial stage of the use of a mHealth app, usability assessment is the key, which can remarkably motivate individuals to use a mHealth app more easily and efficiently.

The definition of usability by the International Organization for Standardization (IOS) :the extent to which a system, product or service can be used by specified users to achieve specified goals with effectiveness, efficiency and satisfaction in a specified circumstance (International Organization for Standardization [ISO], 2018). To evaluate the usability of apps, qualitative methods are used previously, such as interviews (Stinson et al., 2006), cognitive walkthrough (Van Velsen et al., 2018) and heuristic evaluation (Walsh et al., 2017), etc. However, these methods are mainly for tracking usability problems. They do not calculate the absolute score of a system’s usability, which can be achieved via usability tests instead. A usability test is a clear indicator to show whether the usability of an app is sufficient or insufficient (Broekhuis et al., 2021). Ideally, usability test should be conducted at each step of the program development, which includes an iterative cycle of system design and validation (Ivory and Marti, 2001). Testing the usability of an app is most frequently done by means of usability questionnaires (Peute et al., 2008).

The usability Questionnaire is the most common tool for usability assessment, as it is user-friendly and the data analysis is intuitive (Sevilla-Gonzalez et al., 2020). There are currently many usability questionnaires that are validated and reliable, such as Questionnaire for User Interface Satisfaction (Chin et al., 1988),the Post-Study System Usability Questionnaire (PSSUQ) (Lewis, 2002), Perceived Usefulness and Ease of Use (Davis, 1989), System Usability Scale (SUS) (Brooke, 1996), and Usefulness, Satisfaction, and Ease of Use Questionnaire (Lund, 2001). These questionnaires have been used widely in lots of mHealth app usability studies. However, none of them was specifically designed for evaluating the usability of mHealth apps (Zhou et al., 2017). Despite the large number of mHealth apps released to the public, there is still a lack of targeted evaluation tools. Therefore, Zhou et al. (2019) developed and validated a new mHealth app usability questionnaire (MAUQ), which is a targeted reliable usability testing tool in the mHealth domain. The MAUQ, based on the heterogeneity of definitions and methods, is used to evaluate the usability of mHealth apps. Three usability-based dimensions that were consistent with the definition were explored: Ease of use and satisfaction (corresponding to satisfaction of usability); System information arrangement (corresponding to the efficiency of availability); and practicability (corresponding to the availability).

MHealth apps can be divided into two different versions for patients and healthcare providers depending on the target audience. According to the interactive status of mHealth apps, they can be divided into interactive mHealth apps and standalone mHealth apps. Zhou et al. (2019) developed the MAUQ based on a number of existing questionnaires used in previous mobile app usability studies. Then, they utilized MAUQ, SUS, and PSSUQ to investigate the usability of two mHealth apps: an interactive mHealth app and a standalone mHealth app. Four versions of the MAUQ were developed in association with the type of app (interactive or standalone) and target user of the app (patient or provider). A website was created to make it convenient for developers of mHealth apps to use the MAUQ to assess the usability of their mHealth apps (PITT Usability Questionnaire, 2021).

MHealth app usability questionnaire has been widely utilized in the United States (Sengupta et al., 2020; Sood et al., 2020; Cummins et al., 2021), Sweden (Anderberg et al., 2019), Australia (Menon et al., 2019), Spain (Soriano et al., 2020), South Africa (Gous et al., 2020), and other countries. It is mainly used to study the usability of digital healthcare systems. Its standalone patient version has been translated into the Malaysian version, and it has shown an acceptable practicability (Mustafa et al., 2021). However, China has not yet introduced a Chinese version of the MAUQ, the present study aimed to translate and verify the patient version of the interactive mHealth app usability questionnaire (MAUQ) into Chinese, and to provide a usability testing tool for developers of mHealth apps in China.

Materials and Methods

Overview

Zhou et al. (2019) developed the MAUQ in 2019. It contains a total of 21 items and 3 dimensions: usability and satisfaction (8 items), system information arrangement (6 items), and efficiency (7 items). Cronbach’s alpha coefficients of the three dimensions of the original questionnaire were 0.895, 0.829, and 0.900, respectively, indicating a strong internal consistency of the MAUQ. The 7-point Likert scoring system was adopted as follows: 1 (extremely strongly agree), 2 (strongly agree), 3 (agree), 4 (neutral), 5 (disagree), 6 (strongly disagree), and 7 (extremely strongly disagree). The questionnaire score is the total score of each item divided by the number of items. The closer the mean value is to 1, the higher the app’s usability will be.

Questionnaire Translation

It is important to contact a questionnaire’s developer(s) to obtain the permission. The author of the original questionnaire has been contacted and permission for translation and use has been obtained. The Brislin’s translation model (Brislin, 1976) was adopted as follows: Step 1. Forward translation (from English into Chinese): The questionnaire was translated by two translators with high English proficiency, and then, the research group and the two translators could jointly form the Chinese version of the MAUQ (A); Step 2. Back translation: Two translators who had not contacted the original questionnaire translated the Chinese version of the MAUQ (A) into English version, and the research group finalized the English version. After that, the research group and all the translators compared the translated version (From Chinese into English) with the original questionnaire, and the Chinese version of MAUQ (B) could be provided after discussing and performing the required amendments.

Cross-Cultural Adaptation

Literal translation is not sufficient to produce an equivalent questionnaire, and cultural sensitivities or cultural influences may not cause problems, while they may lead to a misunderstanding of the question being asked, indicating the necessity of conducting cross-cultural adaptation (Epstein et al., 2015; Mohamad Marzuki et al., 2018). Therefore, we conducted cross-cultural adaptation according to experts’ comments and prediction test.

Experts’ Comments

Seven experts familiar with the field of mHealth were invited: two associate professors of medicine, one senior engineer of software design, one doctoral assistant researcher, one associate professor of nursing, and two master’s degree nurses. Two researchers of the group took back the expert opinions on site. According to their theoretical knowledge and practical experience, each expert evaluates the accuracy of translation, content comprehension, language expression habits and consistency on cultural background of each item one by one. For the suggestions that are not easy to understand, the researcher had an in-depth discussion with the experts and made records in detail. Finally, the group summarized all the opinions from experts.

Prediction Test

A preliminary survey was conducted on 30 patients who used the mHealth app of “Good Doctor” (“Good Doctor” is the most popular mHealth app in China, and this app showed a high utilization rate), using convenience sampling method, among outpatients of a third-class hospital, aiming to figure out respondents’ understanding of the questionnaire and their feelings on items of the questionnaire. The researcher explained the purpose of the survey to the respondents, made an investigation using the translated questionnaire, and asked the following questions: (1). Do you understand the content of this item? Or whether there is any ambiguity? (2). Do you know how to answer? If you don’t, what might be the difficulty? (3). Does it conform to Chinese language expression habits, if not, in which way you think it should be expressed? After the investigation and communication with the respondents, the items in the questionnaire that can be confusing and difficult to understand or answer were marked. The questions and suggestions raised by the respondents were recorded.

The Chinese version of MAUQ (B) was revised to form the final Chinese version of the MAUQ based on experts’ suggestions and respondents’ feedback.

Research Objects and Research Tools

This cross-sectional, descriptive study was conducted in four large hospitals of Jinan, Shandong Province, China, from October 2020 to February 2021. The convenience sampling method was adopted to select users of the patient version of the “Good Doctor” mHealth app as research objects. Inclusion criteria were as follows: owning a smart phone and being proficient in mobile phone apps; users who aged ≥ 18, ≤ 65 years old, with good verbal and written communication skills; users who have used this app for diagnosis/treatment twice or more in the past month (Sevilla-Gonzalez et al., 2020). Exclusion criteria were as follows: language barriers or communication difficulties; being concomitant with psychiatric or neurological disorders. All participants gave informed consent and voluntarily participated in the study.

After obtaining the study approval from relevant departments, the research team members, who had received unified training, conducted the investigation. The research tools included users’ general information, involving users’ age, gender, place of residence, marital status, educational level, annual income, and the frequency of using “Good Doctor” mHealth app in the last 2 months. The purpose and significance of the questionnaire, as well as the confidentiality principle were explained to subjects before commencing the survey.

Two well-trained and eligible research investigators collected data. The collected questionnaires are numbered uniformly, and the data was entered by two people, excel used for data entry and SPSS 21 used for statistical analysis, and AMOS 23 statistical software used for confirmatory factor analysis to verify the appropriateness and stability of the construct validity of the scale.

Verification of the Questionnaire

Item Analysis

Examine the central tendency of the answers to an item, if there is a choice in the answer set for an item and the percentage is more than 80%, indicating that the discrimination ability of the item is weak, and should be deleted. The total score of the questionnaire was arranged in a logical order. The t-test was used to examine the mean difference, and items with no significant difference between high-score and low-score groups were deleted. Using the Pearson’s correlation coefficient method, and according to the item score and the total score of the questionnaire, items with a very poor correlation with the total score of the questionnaire were deleted (r < 0.30).

Reliability and Validity Testing of the Questionnaire

Seven experts in mHealth were invited to evaluate the relevance of the questionnaire contents. The 4-point Likert scoring system was used to evaluate the correlation between each item and the questionnaire topic [1 (no correlation), 2 (poor correlation), 3 (strong correlation), and 4 (extremely strong correlation)]. The content validity index (CVI) of all items and the CVI of the total score of the questionnaire were calculated. Principal component analysis and maximum variance orthogonal rotation were utilized for exploratory factor analysis (EFA). Items with a load factor of < 0.40 were deleted. The correlation matrix between each dimension and the total score of the questionnaire was tested and the internal correlation analysis was carried out. Confirmatory factor analysis (CFA) was undertaken to verify whether the structure of the questionnaire was consistent with the theoretical structure of the original questionnaire. The reliability of the questionnaire was evaluated by the Cronbach’s α coefficient, test-retest reliability, and split-half reliability.

This study was approved by the Ethics Committee of Qilu Hospital of Shandong University (Jinan, China; Approval No. KYLL-2020-460).

Results

Cross-Cultural Adaptation

According to the experts’ comments and respondents’ feedback, related to content understanding, language expression habits, and cultural background of the questionnaire, and the results of prediction test, the research group modified the contents as follows: (1). The term “various social environments” in item 5 was not consistent with the language expression habits in China. Therefore, item 5 “I feel comfortable using this app in various social environments” was changed to “I feel comfortable using this app in various public places” (2). Item 6 “The time required to use the app is appropriate for me.” was changed to “Using this app will not take up much of my time” (3). The 12th item was not easy to understand after literal translation, thus, the item 12 “navigation is consistent when switching pages” was changed to “the way and process of switching pages are consistent”.

General Information of the Subjects

A total of 444 electronic questionnaires were forwarded to users of the patient version of the “Good Doctor” mHealth app, and 346 questionnaires were effectively received with an effective recovery rate of 77.920%. Analysis of the results showed that, there were 186 and 160 male and female cases, accounting for 53.757% and 46.243% of the total sample size, respectively. In terms of age, the majority of users aged between 29 and 48 years old. The dominancy of bachelor’s degree was found among users. Besides, users mainly lived in urban areas. The proportion of “in-service” personnel was the highest. Table 1 summarizes further details.

TABLE 1
www.frontiersin.org

Table 1. Subjects’ demographic characteristics (n = 346).

Item Analysis

Table 2 shows the results of the standalone-samples t-test, comparing differences in each item between high-score and low-score groups. Items with no statistically significant difference between high-score and low-score groups were deleted (Jones et al., 2001), and the results showed that significant differences existed in each item between high-score and low-score groups, thus, no item was deleted.

TABLE 2
www.frontiersin.org

Table 2. Item analysis (standalone-samples t-test).

The Pearson’s correlation coefficient was used to assess the correlation among 21 items. As shown in Table 3, all the items were significant, and a positive correlation was found among the values of assessment phase. The correlation coefficient between the score of each item and the total score of the questionnaire was calculated, 0.480∼0.681 (P < 0.001), thus, all items were retained.

TABLE 3
www.frontiersin.org

Table 3. The Pearson’s correlation coefficient between each item and the total score of the questionnaire.

Validity Testing of the Questionnaire

Content Validity

As shown in Table 4, the CVI of each item in the questionnaire was > 0.790, indicated that the items were comprehensible to the target users (Zamanzadeh et al., 2015). and the CVI of the total score of the questionnaire was 0.952.

TABLE 4
www.frontiersin.org

Table 4. Content validity of the questionnaire.

Correlation Analysis

Correlation analysis was used to study the correlation between the overall usability and satisfaction of the questionnaire, the arrangement of system information, and the efficiency of the questionnaire. The Pearson’s correlation coefficient was used to indicate the strength of the correlation. The results of correlation analysis revealed that, usability and satisfaction, arrangement of system information, and efficiency of the questionnaire were all significant, and the correlation coefficients between each dimension and the total score were 0.797, 0.711, and 0.760 respectively, indicating the existence of positive correlations among the usability and satisfaction, the arrangement of system information, and the efficiency of the questionnaire (Table 5).

TABLE 5
www.frontiersin.org

Table 5. The correlation analysis of three dimensions using the Pearson’s correlation coefficient.

Structural Validity

Exploratory Factor Analysis

The validity of the questionnaire was assessed through EFA of the values of Kaiser-Meyer-Olkin (KMO), common degree (common factor variance), variance interpretation rate, load factor, and other indicators. The corresponding common degree values of the research items were all higher than 0.400, Kaiser gave the commonly used KMO measurement standard: above 0.9 means it is very suitable; 0.8 means suitable; 0.7 means general; 0.6 or below means not suitable. The KMO value of this study was 0.932 > 0.800, suitable for factor analysis which indicated that the questionnaire had an acceptable validity. The variance interpretation rates of the three factors were 23.65%, 21.76%, and 18.42%, respectively, and the cumulative variance interpretation rate after rotation was 63.830% > 60%, which revealed that all the information could be extracted effectively. After rotating the component matrix for each question in the questionnaire, the load factor of each item in a certain dimension was > 0.500. Therefore, the validity of the questionnaire was noticeable and the questionnaire was efficient (Table 6).

TABLE 6
www.frontiersin.org

Table 6. The results of EFA.

Confirmatory Factor Analysis

Confirmatory factor analysis was used in this study to indicate whether the structure of the questionnaire was consistent with the theoretical structure of the original questionnaire. As shown in Figure 1, when the measurement model was established, the maximum likelihood estimation method was used to estimate the parameters of the model by maximizing the likelihood function. The estimated parameters were load factor, equation error, and error covariance. The most direct parameter that was used to evaluate the goodness of fit of the model. The evaluation of goodness-of-fit the model included the testing of degree of fitting and testing of parameters. In this study, the minimum discrepancy/degrees of freedom (CMIN/DF), normed fit index (NFI), incremental fit index (IFI), Tucker-Lewis index (TLI), comparative fit index (CFI), goodness-of-fit index (GFI), adjusted GFI (AGFI), root mean square error of approximation (RMSEA), and other model fitness indices all met the requirements, indicating that the fitness of the model was good (Table 7).

FIGURE 1
www.frontiersin.org

Figure 1. The results of CFA.

TABLE 7
www.frontiersin.org

Table 7. The evaluation of the goodness-of-fit of the model.

Analysis of Convergent Validity and Discriminant Validity

In this study, the composite reliability (CR) and average variance extracted (AVE) values were used as the evaluation criteria for convergence validity. When the CR value of each factor is > 0.700 and the AVE value is > 0.500, it is generally considered that the convergent validity is good. In addition, when the square root value of AVE of each factor is higher than the correlation coefficient between this factor and other factors, it indicates a high discriminant validity. The results of convergence validity and discriminant validity are presented in Table 8. As shown in Table 8, the mean AVE value of each dimension is greater than 0.5, and the mean CR value is greater than 0.700, indicating that the convergence validity of the questionnaire is noticeable.

TABLE 8
www.frontiersin.org

Table 8. The results of the convergent validity analysis.

According to the results of the discriminant validity analysis (Table 9), the square root of the AVE was greater than the value of the correlation with other factors, thus, the discriminant validity (among internal factors) of each variable was very promising.

TABLE 9
www.frontiersin.org

Table 9. The results of the discriminant validity analysis using the Pearson’s correlation coefficient.

Reliability Testing of the Questionnaire

The Cronbach’s α coefficient of the Chinese version of the MAUQ was 0.912, higher α value suggests greater internal reliability and more than0.700 is acceptable as good internal reliability (Allen et al., 2014). The values of test-retest reliability were 0.869, and the values of split-half reliability were 0.701. The reliability of each dimension of the questionnaire is presented in Table 10.

TABLE 10
www.frontiersin.org

Table 10. Reliability of each dimension of the questionnaire.

Discussion

This study described the process of cross-cultural translation and adaption of the MAUQ questionnaire Semantic and cultural equivalence was achieved between the two versions and the adapted one showed excellent internal consistency and good validity finally.

The development of new questionnaires requires the joint efforts of members of professional research teams, while attaining the objective is costly and time-consuming. Therefore, it is recommended to adapt the established, reliable, and available questionnaires and record their validity in the language used (Mohamad Marzuki et al., 2018). However, due to the linguistic and cultural differences between Chinese and English, literal translation may cause ambiguity in the meaning expressed in a part of the content, hindering the production of an authenticated questionnaire. Therefore, cross-cultural adaptation of the questionnaire is necessary (Epstein et al., 2015). The present study used the Brislin’s translation model (Brislin, 1976) for translation, and the translation results showed that the majority of elements were easy to understand, while few items were adjusted after translation according to Chinese expression habits.

The verification analysis of the questionnaire is an important step to ensure that the composition of the translated version of the questionnaire is as same as that of the original version. CVI and reliability index were used in the current research. The CVI is basically used to measure the validity of the questionnaire (Maramba et al., 2019). It is easy to measure and understand, and the content of each item can be modified or deleted according to the detailed information of each item (Polit et al., 2007; Zamanzadeh et al., 2015). The results of this study showed that the CVI value of each item was 0.857∼1.000, and the total CVI was 0.952 > 0.790 (benchmark value), indicating that the Chinese version of the MAUQ could properly represent each item. Reliability is the degree to which an assessment tool produces stable and consistent results. In the present study, the Cronbach’s α coefficient, split-half reliability, and test-retest reliability were used to evaluate the reliability of the questionnaire. The results showed that the values of Cronbach’s α coefficient, test-retest reliability, and split-half reliability were 0.912 > 0.800, 0.869 > 0.800, and 0.701 > 0.700, respectively. The results confirmed that the questionnaire had a good internal consistency and stability.

In the present study, the reliability and validity of the MAUQ were tested to provide an evaluation tool for mobile medical services in China. However, there are some limitations in this study. Firstly, 346 patients recruited in this study were outpatients from a third-class hospital, particularly youth who were familiar with smart phones. The results of this study have not been generalized to a wider population, especially the elderly. Secondly, this study only concentrated on the evaluation of the patient version of the “Good Doctor” app, thus, the scope of our research was limited, and it could not be applied to other versions (provider version or medical version) of different mHealth apps. In the future research, we will concentrate on the usability testing tools for mHealth apps in all aspects, so as to further promote the development of mHealth apps.

Conclusion

The Chinese version of the MAUQ, an interactive mHealth app for patients, contains 21 items and 3 dimensions, which is consistent with the theoretical structure of the original questionnaire and has a high reliability and validity. All indicators of the Chinese version of the MAUQ could meet the requirements of the measurement, which could effectively and scientifically evaluate the patient version of the interactive mHealth app.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by Qilu Hospital of Shandong University. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

SZ and YC contributed to the study design, implementation, analysis, and manuscript writing. HC and KL contributed to the statistical design and analysis. XL, JZ, and YL contributed to the analysis of the data. PD contributed to the translation of the scale. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We would like to thank Leming Zhou for granting permission to convert the MAUQ into Chinese. We highly appreciate the author PD for his guidance during the translation of the MAUQ. We also acknowledge Chongzhong Liu, Yu Han, and Qianqian Xu for providing significant technical comments on the questionnaire content.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2022.813309/full#supplementary-material

Abbreviations

CVI, content validity index; KMO, Kaiser Meyer Olkin; MAUQ, mHealth app usability questionnaire; mHealth, mobile health.

References

Allen, P., Bennett, K., and Heritage, B. (2014). SPSS Statistics version 22: A Practical Guide. Sydney: Cengage Learning Australia Pty Limited.

Google Scholar

Anderberg, P., Eivazzadeh, S., and Berglund, J. S. (2019). A novel instrument for measuring older people’s attitudes toward technology (TechPH): development and validation. J. Med. Internet Res. 21:e13951. doi: 10.2196/13951

PubMed Abstract | CrossRef Full Text | Google Scholar

Association of American Medical Colleges [AAMC] (2021). The Complexities of Physician Supply and Demand: Projections from 2018 to 2033. Washington, DC: Association of American Medical Colleges.

Google Scholar

Brislin, R. W. (1976). Comparative research methodology: cross-cultural studies. Int. J. Psychol. 113, 215–229. doi: 10.1080/00207597608247359

CrossRef Full Text | Google Scholar

Broekhuis, M., van Velsen, L., Peute, L., Halim, M., and Hermens, H. (2021). Conceptualizing usability for the eHealth context: content analysis of usability problems of eHealth applications. JMIR Form. Res. 5:e18198. doi: 10.2196/18198

PubMed Abstract | CrossRef Full Text | Google Scholar

Brooke, J. (1996). SUS: A Quick and Dirty Usability Scale. Boca Raton, FL: CRC Press, 189–194.

Google Scholar

Chin, J. P., Diehl, V. A., and Norman, K. L. (1988). “Development of an instrument measuring user satisfaction of the human-computer interface,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, (Gaithersburg, MD: ACM Press).

Google Scholar

Cummins, K. M., Brumback, T., Chung, T., Moore, R. C., Henthorn, T., Eberson, S., et al. (2021). Acceptability, validity, and engagement with a mobile app for frequent, continuous multiyear assessment of youth health behaviors (mNCANDA): mixed methods study. JMIR Mhealth Uhealth 9:e24472. doi: 10.2196/24472

PubMed Abstract | CrossRef Full Text | Google Scholar

Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. 13, 319–340. doi: 10.2307/249008

CrossRef Full Text | Google Scholar

Epstein, J., Santo, R. M., and Guillemin, F. (2015). A review of guidelines for cross-cultural adaptation of questionnaires could not bring out a consensus. J. Clin. Epidemiol. 68, 435–441. doi: 10.1016/j.jclinepi.2014.11.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Fairman, A. D., Dicianno, B. E., Datt, N., Garver, A., Parmanto, B., and McCue, M. (2013). Outcomes of clinicians, caregivers, family members and adults with Spina bifida regarding receptivity to use of the iMHere mHealth solution to promote wellness. Int. J. Telerehabil. 5, 3–16. doi: 10.5195/ijt.2013.6116

PubMed Abstract | CrossRef Full Text | Google Scholar

Gous, N., Fischer, A. E., Rhagnath, N., Phatsoane, M., Majam, M., and Lalla-Edward, S. T. (2020). Evaluation of a mobile application to support HIV self-testing in Johannesburg, South Africa. South. Afr. J. HIV Med. 21:1088. doi: 10.4102/sajhivmed.v21i1.1088

PubMed Abstract | CrossRef Full Text | Google Scholar

International Organization for Standardization [ISO] (2018). ISO 9241. International Organization for Standardization. Available Online at: https://www.iso.org/obp/ui/#iso:std:iso:9241:-11:ed-2:v1:en (accessed November 11, 2019).

Google Scholar

Ivory, M. Y., and Marti, A. H. (2001). The state of the art in automating usability evaluation of user interfaces. ACM Comput. Surv. 33, 470–516. doi: 10.1145/503112.503114

CrossRef Full Text | Google Scholar

Jones, P. S., Lee, J. W., Phillips, L. R., Zhang, X. E., and Jaceldo, K. B. (2001). An adaptation of Brislin’s translation model for cross-cultural research. Nurs. Res. 50, 300–304. doi: 10.1097/00006199-200109000-00008

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, J. R. (2002). Psychometric evaluation of the PSSUQ using data from five years of usability studies. Int. J. Hum. Comput. Interact. 14, 463–488. doi: 10.1207/s15327590ijhc143%264_11

CrossRef Full Text | Google Scholar

Lund, A. (2001). STC Usability SIG Newsletter. Measuring Usability with the USE Questionnaire. Available Online at: https://www.researchgate.net/publication/230786746_Measuring_usability_with_the_USE_questionnaire (accessed July 28, 2017).

Google Scholar

Mann, D. M., Chen, J., Chunara, R., Testa, P. A., and Nov, O. (2020). COVID-19 transforms health care through telemedicine: evidence from the field. J. Am. Med. Inform. Assoc. 27, 1132–1135. doi: 10.1093/jamia/ocaa072

PubMed Abstract | CrossRef Full Text | Google Scholar

Maramba, I., Chatterjee, A., and Newman, C. (2019). Methods of usability testing in the development of eHealth applications: a scoping review. Int. J. Med. Inform. 126, 95–104. doi: 10.1016/j.ijmedinf.2019.03.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Menon, A., Fatehi, F., Ding, H., Bird, D., Karunanithi, M., Gray, L., et al. (2019). Outcomes of a feasibility trial using an innovative mobile health programme to assist in insulin dose adjustment. BMJ Health Care Inform. 26:e100068. doi: 10.1136/bmjhci-2019-100068

PubMed Abstract | CrossRef Full Text | Google Scholar

Mohamad Marzuki, M. F., Yaacob, N. A., and Yaacob, N. M. (2018). Translation, cross-cultural adaptation, and validation of the Malay version of the system usability scale questionnaire for the assessment of mobile apps. JMIR Hum. Factors 5:e10308. doi: 10.2196/10308

PubMed Abstract | CrossRef Full Text | Google Scholar

Mustafa, N., Safii, N. S., Jaffar, A., Sani, N. S., Mohamad, M. I., Rahman, A. H. A., et al. (2021). Malay version of the mHealth app usability questionnaire (M-MAUQ): translation, adaptation, and validation study. JMIR Mhealth Uhealth 9:e24457. doi: 10.2196/24457

PubMed Abstract | CrossRef Full Text | Google Scholar

National Health Service [NHS] (2019). The NHS Long Term Plan. London: NHS.

Google Scholar

Parmanto, B., Pramana, G., Yu, D. X., Fairman, A. D., Dicianno, B. E., and McCue, M. P. (2013). iMHere: a novel mHealth system for supporting self-care in management of complex and chronic conditions. JMIR Mhealth Uhealth 1:e10. doi: 10.2196/mhealth.2391

PubMed Abstract | CrossRef Full Text | Google Scholar

Peute, L. W. P., Spithoven, R., Bakker, P. J. M., and Jaspers, M. W. M. (2008). Usability studies on interactive health information systems; where do we stand? Stud. Health Technol. Inform. 136, 327–332.

Google Scholar

Pfammatter, A., Spring, B., Saligram, N., Davé, R., Gowda, A., Blais, L., et al. (2016). mHealth intervention to improve diabetes risk behaviors in India: a prospective, parallel group cohort study. J. Med. Internet Res. 18:e207. doi: 10.2196/jmir.5712

PubMed Abstract | CrossRef Full Text | Google Scholar

PITT Usability Questionnaire (2021). For Telehealth Systems (TUQ) and Mobile Health Apps (MAUQ). Available Online at: https://ux.hari.pitt.edu/app/about (accessed July 11, 2021).

Google Scholar

Polit, D. F., Beck, C. T., and Owen, S. V. (2007). Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Res. Nurs. Health 30, 459–467. doi: 10.1002/nur.20199

PubMed Abstract | CrossRef Full Text | Google Scholar

Schinkothe, T., Gabri, M. R., Mitterer, M., Gouveia, P., Heinemann, V., Harbeck, N., et al. (2020). A web- and app-based connected care solution for COVID-19 in- and outpatient care: qualitative study and application development. JMIR Public Health Surveill. 6:e19033. doi: 10.2196/19033

PubMed Abstract | CrossRef Full Text | Google Scholar

Sengupta, A., Beckie, T., Dutta, K., Dey, A., and Chellappan, S. (2020). A mobile health intervention system for women with coronary heart disease: usability study. JMIR Form. Res. 4:e16420. doi: 10.2196/16420

PubMed Abstract | CrossRef Full Text | Google Scholar

Seto, E., Leonard, K. J., Cafazzo, J. A., Barnsley, J., Masino, C., and Ross, H. J. (2012). Perceptions and experiences of heart failure patients and clinicians on the use of mobile phone-based telemonitoring. J. Med. Internet Res. 14:e25. doi: 10.2196/jmir.1912

PubMed Abstract | CrossRef Full Text | Google Scholar

Sevilla-Gonzalez, M. D. R., Moreno Loaeza, L., Lazaro-Carrera, L. S., Ramirez, B. B., Rodríguez, A. V., Peralta-Pedrero, M. L., et al. (2020). Spanish version of the system usability scale for the assessment of electronic tools: development and validation. JMIR Hum. Factors. 7:e21161. doi: 10.2196/21161

PubMed Abstract | CrossRef Full Text | Google Scholar

Shah, S. S., Gvozdanovic, A., Knight, M., and Gagnon, J. (2021). Mobile app-based remote patient monitoring in acute medical conditions: prospective feasibility study exploring digital health solutions on clinical workload during the COVID crisis. JMIR Form. Res. 5:e23190. doi: 10.2196/23190

PubMed Abstract | CrossRef Full Text | Google Scholar

Sood, R., Stoehr, J. R., Janes, L. E., Ko, J. H., Dumanian, G. A., and Jordan, S. W. (2020). Cell phone application to monitor pain and quality of life in neurogenic pain patients. Plast. Reconstr. Surg. Glob. Open 8:e2732. doi: 10.1097/GOX.0000000000002732

PubMed Abstract | CrossRef Full Text | Google Scholar

Soriano, J. B., Fernandez, E., de Astorza, A., de Llano, L. A. P., Fernández-Villar, A., Carnicer-Pont, D., et al. (2020). Hospital epidemics tracker (HEpiTracker): description and pilot study of a mobile app to track COVID-19 in hospital workers. JMIR Public Health Surveill. 6:e21653. doi: 10.2196/21653

PubMed Abstract | CrossRef Full Text | Google Scholar

Stinson, J. N., Petroz, G. C., Tait, G., Feldman, B. M., Streiner, D., McGrath, P. J., et al. (2006). e-Ouch: usability testing of an electronic chronic pain diary for adolescents with arthritis. Clin. J. Pain 22, 295–305. doi: 10.1097/01.ajp.0000173371.54579.31

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Velsen, L., Evers, M., Bara, C.-D., Den Akker, H. O., Boerema, S., and Hermens, H. (2018). Understanding the acceptance of an eHealth technology in the early stages of development: an end-user walkthrough approach and two case studies. JMIR Form. Res. 2:e10474. doi: 10.2196/10474

PubMed Abstract | CrossRef Full Text | Google Scholar

Vera, F., Noel, R., and Taramasco, C. (2019). Standards, processes and instruments for assessing usability of health mobile apps: a systematic literature review. Stud. Health Technol. Inform. 264, 1797–1798. doi: 10.3233/SHTI190653

PubMed Abstract | CrossRef Full Text | Google Scholar

Walsh, L., Hemsley, B., Allan, M., Adams, N., Balandin, S., Georgiou, A., et al. (2017). The e-health literacy demands of Australia’s my health record: a heuristic evaluation of usability. Perspect. Health Inf. Manag. 14:1f.

Google Scholar

World Health Organization [WHO] (2020). Process of Translation and Adaptation of Instruments. Geneva: World Health Organization.

Google Scholar

Zamanzadeh, V., Ghahramanian, A., Rassouli, M., Abbaszadeh, A., Alavi-Majd, H., and Nikanfar, A. R. (2015). Design and implementation content validity study: development of an instrument for measuring patient-centered communication. J. Caring Sci. 4, 165–178. doi: 10.15171/jcs.2015.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, L., Bao, J., and Parmanto, B. (2017). Systematic review protocol to assess the effectiveness of usability questionnaires in mHealth app studies. JMIR Res. Protoc. 6:e151. doi: 10.2196/resprot.7826

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, L., Bao, J., Setiawan, I. M. A., Saptono, A., and Parmanto, B. (2019). The mHealth app usability questionnaire (MAUQ): development and validation study. JMIR Mhealth Uhealth 7:e11500. doi: 10.2196/11500

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: mHealth apps, content validity index, cross-cultural adaptation, questionnaire translation, usability testing tools

Citation: Zhao SQ, Cao YJ, Cao H, Liu K, Lv XY, Zhang JX, Li YX and Davidson PM (2022) Chinese Version of the mHealth App Usability Questionnaire: Cross-Cultural Adaptation and Validation. Front. Psychol. 13:813309. doi: 10.3389/fpsyg.2022.813309

Received: 03 December 2021; Accepted: 10 January 2022;
Published: 02 February 2022.

Edited by:

Sina Hafizi, University of Manitoba, Canada

Reviewed by:

Donna Slovensky, University of Alabama at Birmingham, United States
Shuo Zhou, University of Colorado, Denver, United States

Copyright © 2022 Zhao, Cao, Cao, Liu, Lv, Zhang, Li and Davidson. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yingjuan Cao, caoyingjuan@hotmail.com; caoyj@sdu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.