Abstract
Background
An artificial intelligence (AI)–assisted contouring system benefits radiation oncologists by saving time and improving treatment accuracy. Yet, there is much hope and fear surrounding such technologies, and this fear can manifest as resistance from health care professionals, which can lead to the failure of AI projects.
Objective
The objective of this study was to develop and test a model for investigating the factors that drive radiation oncologists’ acceptance of AI contouring technology in a Chinese context.
Methods
A model of AI-assisted contouring technology acceptance was developed based on the Unified Theory of Acceptance and Use of Technology (UTAUT) model by adding the variables of perceived risk and resistance that were proposed in this study. The model included 8 constructs with 29 questionnaire items. A total of 307 respondents completed the questionnaires. Structural equation modeling was conducted to evaluate the model’s path effects, significance, and fitness.
Results
The overall fitness indices for the model were evaluated and showed that the model was a good fit to the data. Behavioral intention was significantly affected by performance expectancy (β=.155; P=.01), social influence (β=.365; P<.001), and facilitating conditions (β=.459; P<.001). Effort expectancy (β=.055; P=.45), perceived risk (β=−.048; P=.35), and resistance bias (β=−.020; P=.63) did not significantly affect behavioral intention.
Conclusions
The physicians’ overall perceptions of an AI-assisted technology for radiation contouring were high. Technology resistance among Chinese radiation oncologists was low and not related to behavioral intention. Not all of the factors in the Venkatesh UTAUT model applied to AI technology adoption among physicians in a Chinese context.
Keywords: artificial intelligence, technology acceptance model, intension, resistance
Introduction
Background
The artificial intelligence (AI) race has long since begun; its silicone roots date back to 1956 [1]. In China’s “Development Plan for The New Generation of Artificial Intelligence,” the goal was to synchronize China’s overall technology and AI applications with the world's advanced level of technology by 2020. The total value of China’s AI industry market is expected to reach US $127 billion by 2025 [2], and the medical AI industry is forecasted to account for one-fifth of the overall AI market [3]. In the health care industry, AI has been developed for administrative assistance, clinical decision-making support, patient monitoring, and robotic surgeries [4]. In clinical decision-making support, computer software is used to analyze clinical data and assist medical professionals in making more precise decisions on disease diagnoses and treatments [5]. Sketching tumors is one such subfield.
Delineating tumors is tricky and requires much education, guidance, and expertise. For radiation oncologists, the tumor target contouring process for precision head and neck radiation therapy is time consuming. It requires, on average, 180 minutes of intense work for each patient, and contouring areas that are created by different radiation oncologists for the same tumor can be highly variable [6]. Moreover, contouring accuracy is key for achieving successful treatment outcomes. Suboptimal tumor coverage and poor-quality radiation therapy plans are major factors of disease relapse and inferior survival [7]. An AI contouring tool that automates the primary gross tumor volume contouring process for patients with cancer was developed at Sun Yat-sen University Cancer Center, wherein it has been applied clinically since 2019; over 12,500 cases of radiation planning have been carried out with the tool’s assistance. The contouring system has a high level of accuracy (ie, when compared with the manual contouring accuracy of radiation oncologists). Although oncologists still need to edit automated contours by making minor adjustments, the system substantially saves time by up to 39.4%, and intraobserver and interobserver variations have been reduced by 36.4% and 54.5%, respectively [8].
Although the benefits of AI assistance appear to be obvious, like all forms of change, the adoption of AI technology might be met with resistance, especially within health care settings; it has been said that the only constant is change [9]. There is much hope and fear surrounding such technologies, and this fear can manifest as resistance from both health care professionals and patients [10]. One particular fear that may promote resistance is the fear that AI tools will replace workers (even skilled workers like doctors)—the much discredited, but still prevalent, Luddite fallacy [11]. In a British Medical Journal head-to-head debate, the topic was as follows: “Could machines using artificial intelligence make doctors obsolete?” Dr Jörg Goldhahn said “yes” and foreboded that humans simply cannot keep up with the data crunching advantage that computers have over humans [12]. Further, Vanessa Ramptom and Professor Giatgen Spinas claimed that the role of doctors will not become obsolete; instead, these roles will change to specifically focus on treating a patient as a whole [12]. These perspectives encompass part of the larger narratives that describe AI technology as either utopian or dystopian.
There are also claims that doctors are slow to uptake new technologies. In 2004, there was evidence of resistance across the United States against using a computerized order entry system to overcome issues, such as doctors’ illegible handwriting, and reduce the number of incorrect prescriptions. The Cedars-Sinai Medical Center notably forced the withdrawal of the system, which was installed in two-thirds of the 870-bed hospital [13]. Similarly, in 2006, Bhattacherjee and Hikmet [14,15] found that physicians were resistant to adopting administrative health care information technologies, and more recently, in 2014, Hsieh [16] discovered 5 factors (attitude, subjective norm, perceived behavior control, institutional trust, and perceived risk) that influenced the adoption of such technologies. Notably, these studies all concerned administrative technologies. There is, however, also evidence of doctors resisting AI technology for other uses, such as patient diagnosis and treatment plan development.
Even though the National Institute of Health claims that AI technology is trending in cancer care [17], a number of notable AI projects have failed in the United States. The MD Anderson Cancer Center, one of the top cancer centers in the world, cancelled a project that involved IBM Corporation’s oncology expert advisor tool, Watson, in 2016 after spending US $62 million [18]. In February 2021, IBM Corporation also tried to sell their Watson Health business, as it was not profitable [19].
In this study, we assessed whether there has been any resistance to the uptake of an AI contouring system in China and examined the system’s pros and cons. Few prior studies have focused on the technology acceptance of AI tools that provide clinical decision-making support in cancer care from physicians’ perspectives. AI health care researchers have primarily analyzed technical innovation and clinical results but have seldom considered the human contexts that are invariably involved in use behavior. Social and implemental issues need to be addressed when an AI tool is introduced in hospitals. Further, it is especially important to examine the extent of perceived risks and technology resistance bias.
Objective
The objective of this study was to develop and test a model for investigating the factors that drive Chinese radiation oncologists’ acceptance of an AI contouring tool for automating the primary gross tumor volume and normal tissue contouring process. This study analyzes how expectancy, social influence, facilitating conditions, perceived risk, and resistance bias contribute to the Technology Acceptance Model (TAM).
Theoretical Background and Hypothesis Development
The theoretical models in this study included the TAM and the Unified Theory of Acceptance and Use of Technology (UTAUT) model. The perceived risk and status quo bias theory variables were added to the UTAUT model.
The first TAM was developed by Fred D Davis in 1989, and it was based on the theory of reasoned action. The TAM was used to analyze consumers’ behaviors in the context of health information technology based on the following two major variables: perceived usefulness and perceived ease of use [20]. Later, in 2000, Venkatesh and Davis [21] developed the TAM2, which also addressed perceived usefulness and use intentions in terms of social influence and cognitive instrumental processes. They added the subjective norm, voluntariness, image, job relevance, output quality, and result demonstrability constructs, which significantly influenced user acceptance. The evolution of the TAM3 came next; the TAM3 was a comprehensive nomological network (ie, an integrated model) for evaluating individual-level (information technology) adoption and use [22].
In 2003, Venkatesh et al [23] also created a UTAUT model, which is an integrated tool based on 8 original models that were published in major information systems journals from a range of disciplines. These models included the theory of reasoned action model, the TAM, the motivation model, the theory of planned behavior model, a combination of the TAM and the theory of planned behavior model, the model of PC utilization, the innovation diffusion theory model, and the social cognitive theory model. The UTAUT originally explained technology acceptance among employees, and in 2012, a UTAUT2 model was developed, which focused on consumer technologies [24]. As AI-assisted contouring technology has been developed for medical staff in hospitals, hedonic motivation and price value, which were explored in the UTAUT2 model, were considered inapplicable to explaining oncologists’ acceptance behaviors in this study. Therefore, we used the original UTAUT model in this study. The original model and its various modifications have been popularly used in studies of technology adoption in the health care industry over the past 5 years [25-28].
The UTAUT model revealed the following four variables: performance expectancy, effort expectancy, social influence (which is used to evaluate the behavioral intention to use a new technology), and facilitating conditions (which are used to evaluate use behavior). Most research has proven that these four constructs positively affect potential users’ behavioral intentions [29,30]. Performance expectancy is defined as “the degree to which an individual believes that using a new technology will help him or her to attain gains in job performance.” This construct was integrated from a similar concept in the TAM—perceived usefulness [31]. Effort expectancy is defined as “the degree of ease of use associated with IT” [23]. This concept is similar to the perceived ease of use construct in the TAM and innovation diffusion theory and the complexity of technology construct in the model of PC utilization. Social influence is defined as “the degree to which an individual perceives the importance of how others believe that he or she should use new IT.” This concept was developed based on the notion that “an individual’s behavior is influenced by the way in which one believes others will view him/her as a result of having used health IT” [23]. The Venkatesh model integrated subjective norms from the theory of reasoned action, the TAM2, and the theory of planned behavior; social factors from the model of PC utilization; and the image construct from the innovation diffusion theory into the social influence construct. The facilitating conditions construct is defined as “the degree to which an individual believes that an organizational and technical infrastructure exists to support the use of the system” [23]. The UTAUT’s facilitating conditions concept contains the perceived behavior control concept from the theory of planned behavior and the combined TAM and theory of planned behavior model, the facilitating conditions concept from the model of PC utilization, and the compatibility construct from the innovation diffusion theory [23]. Lastly, the intention-usage behavior relationship has long been discussed in many studies of technology acceptance, and we expected that behavioral intentions to use AI-assisted technology would have a significant influence on use behavior. Based on these variables, we proposed the following hypotheses:
Hypothesis 1: Performance expectancy positively affects oncologists’ behavioral intentions to use an AI contouring tool for developing radiation therapy plans.
Hypothesis 2: Effort expectancy positively affects oncologists’ behavioral intentions to use an AI contouring tool for developing radiation therapy plans.
Hypothesis 3: Social influence positively affects oncologists’ behavioral intentions to use an AI contouring tool for developing radiation therapy plans.
Hypothesis 4: Oncologists’ behavioral intentions positively affect their use behaviors when using an AI contouring tool for developing radiation therapy plans.
Hypothesis 5A: Facilitating conditions positively affect oncologists’ use behavior when using an AI contouring tool for developing radiation therapy plans.
Hypothesis 5B: Facilitating conditions positively affect oncologists’ behavioral intentions to use an AI contouring tool for developing radiation therapy plans.
Perceived risk has long been discussed as a negative factor in health care information technology adoption studies [10,16], and it refers to the combination of uncertainty and the seriousness of an outcome [32]. Featherman and Pavlou [33] identified the following seven facets of e-service–related perceived risk: performance risk, financial risk, time risk, physiological risk, social risk, privacy risk, and overall risk. In this study, we addressed perceived risk based on the following four aspects: (1) performance risk, which we defined as the probability of malfunction and performance failure and whether the AI contouring tool could mislead physicians with inaccurate contouring; (2) time risk, which we defined as the probability of physicians needing more time to correct errors caused by the AI system; (3) psychological risk, which we analyzed in terms of whether using the AI contouring tool may cause psychological discomfort due to having a negative effect on clinicians’ self-perceptions of the treatment plan; and (4) privacy risk, which we defined as a potential loss of the confidential patient data in the AI-assisted system. As the AI systems were purchased by public hospitals without input from individual physicians, no financial or social risk assessment was required.
Resistance to change refers to people’s attempts to maintain previous behaviors or habits that are connected to their past experiences when facing changes [15,34,35], and this has been discussed as a major barrier to health care information technology adoption [15,36]. The fear of job loss has also been discussed as a factor of the slow adoption of new information technologies [37]. We suggest that radiation oncologists might believe that AI-assisted technologies might replace them and that the adoption of such technologies will lead to an increase in unemployment rates. Thus, we included these two factors in the resistance bias construct, which we defined as people’s resistance to using a new technology resulting from their resistance to change and concerns about unemployment. Therefore, we proposed the following hypotheses:
Hypothesis 6: Perceived risk negatively affects oncologists’ behavioral intentions to use an AI contouring tool for developing radiation therapy plans.
Hypothesis 7: Resistance bias negatively affects oncologists’ behavioral intentions to use an AI contouring tool for developing radiation therapy plans.
Methods
Participants and Sampling
The majority of participants were radiation oncologists (227/307, 73.9%) and medical students (80/307, 26.1%) from the Department of Radiation Oncology at Sun Yat-sen University Cancer Center who had clinical experience in using the computational system for contouring. The sample pool also included radiation oncologists from other hospitals in China. Participants were recruited between July 2020 and November 2020. A convenience sampling method was used in this study, as only a few hospitals in China have AI-assisted contouring systems. As our survey had 29 items, the required number of participants was more than 290. The questionnaire was conducted by sending direct website links via email or WeChat (a social media platform developed by Tencent Holding Limited). Questionnaires were only considered valid if (1) an account responded only once and (2) the total response time for completing the questionnaire was more than 60 seconds. In total, 322 radiation oncologists replied to the web-based survey. After filtering out the invalid responses, 307 responses—176 from Sun Yat-sen University Cancer Center and 131 from other hospitals—were deemed valid for data analysis.
Measurements
Our model included 8 constructs with 29 questionnaire items. Each item assessed 1 construct and was sourced from relevant literature related to new technology acceptance research. Some changes were made to fit the AI-assisted contouring system context (Table 1). After we developed a list of questions, a team discussion was held to reach a consensus. Afterward, we asked a group of 10 oncologists with experience in working with the AI system to answer the questions and provide comments for system improvement. A 7-point Likert scale, which ranged from 1 (strongly disagree) to 7 (strongly agree), was used to assess the items for all constructs except those for the use behavior construct, in which the answers were “yes” or “no.”
Table 1.
Variables and items | Questionnaire items | Sources | |
PEa,b | |||
|
PE1 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
PE2 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
PE3 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
PE4 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
EEa,e | |||
|
EE1 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
EE2 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
EE3 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
EE4 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
SIa,f | |||
|
SI1 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
SI2 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
SI3 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
SI4 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
FCa,g | |||
|
FC1 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
FC2 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
FC3 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
FC4 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
BIa,h | |||
|
BI1 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
BI2 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
|
BI3 |
|
Venkatesh et al [23] and Kijsanayotin et al [25] |
PRa,i | |||
|
PR1 |
|
Ye et al [10], Bhattacherjee and Hikmet [15], and Andrews et al [38] |
|
PR2 |
|
Ye et al [10], Bhattacherjee and Hikmet [15], and Andrews et al [38] |
|
PR3 |
|
Ye et al [10], Bhattacherjee and Hikmet [15], and Andrews et al [38] |
|
PR4 |
|
Ye et al [10], Bhattacherjee and Hikmet [15], and Andrews et al [38] |
RBa,j | |||
|
RB1 |
|
Bhattacherjee and Hikmet [15] and Dou et al [34] |
|
RB2 |
|
Bhattacherjee and Hikmet [15] and Dou et al [34] |
|
RB3 |
|
Raza et al [37] |
UBk,l | |||
|
UB1 |
|
Venkatesh et al [23] |
|
UB2 |
|
Venkatesh et al [23] |
|
UB3 |
|
Venkatesh et al [23] |
a7-point scale (1=strongly disagree; 2=quite disagree; 3=slightly disagree; 4=neither agree or disagree; 5=slightly agree; 6=quite agree; 7=strongly agree).
bPE: performance expectancy.
cQuestions for those who have already used the system.
dQuestions for those who have not used the system.
eEE: effort expectancy.
fSI: social influence.
gFC: facilitating conditions.
hBI: behavioral intention.
iPR: perceived risk.
jRB: resistance bias.
k2-point scale (1=yes; 2=no).
lUB: use behavior.
The first page of the questionnaire provided an overview of this study’s background, purpose, voluntary nature, and anonymity and an introduction of the AI-assisted tumor volume contouring system, which included information on its general functions and operating procedures and images of the system to help the participants gain a clear understanding. We developed 2 questionnaires to assess oncologists’ perceptions. All participants first answered the following question: “Have you used the AI-assisted contouring system?” If the answer was “yes,” then they were directed to answer the questions for those who have already used the system (Table 1); if the answer was “no,” then they were directed to answer the questions for those who have not used the system. The questionnaires were identical apart from the usage of verb tenses. The questions were translated from English to Mandarin Chinese for the survey. Ethical approval was obtained from the Ethics Committee of the Department of Anthropology at Sun Yat-sen University.
Data Analysis
Structural equation modeling analysis was conducted by using IBM SPSS Amos 21.0 (IBM Corporation), and this was a 2-step analysis [39]. First, items and constructs were evaluated for reliability and validity via a confirmatory factor analysis of the measurement model. Second, the structural model’s path effects, significance, and goodness of fit and mediation effects were evaluated. More specifically, the path analysis was performed by conducting a series of multiple regression analyses to derive the standardized regression coefficients (ie, β) for all proposed paths. Figure 1 shows the conceptual model for the analysis; the arrows with solid lines are the paths indicating the relationships among the specified concepts in the hypothesis. IBM SPSS Version 25.0 (IBM Corporation) was used for the descriptive statistics analysis.
Results
Demographic Results
The demographic characteristics of the 307 participants are shown in Table 2. About an equal number of men (151/307, 49.2%) and women (156/307, 50.8%) participated in this study. The majority (269/307, 87.6%) of the participants’ ages ranged between 18 and 40 years, and all of the participants had at least a bachelor’s degree. In China, radiation oncologists who work on contouring are often aged under 40 years, as senior oncologists mainly focus on checking the accuracy of contouring based on the treatment plans, which are mostly developed by junior oncologists. Thus, it was unsurprising that we had more participants aged under 40 years. In our study, 185 participants had at least 4 years of experience in radiation oncology, 123 participants had already used the AI-assisted contouring technology, and 184 had not yet used it.
Table 2.
Characteristics | Participants, n (%) | |
Sex | ||
|
Male | 151 (49.2) |
|
Female | 156 (50.8) |
Age (years) | ||
|
18-25 | 121 (39.4) |
|
26-30 | 67 (21.8) |
|
31-40 | 81 (26.4) |
|
41-50 | 30 (9.8) |
|
>50 | 8 (2.6) |
Education | ||
|
Bachelor’s degree | 137 (44.6) |
|
Master’s degree or doctoral degree | 170 (55.4) |
Experience in radiation oncology (years) | ||
|
≤3 | 122 (39.7) |
|
4-6 | 61 (19.9) |
|
7-10 | 23 (7.5) |
|
>10 | 101 (32.9) |
Artificial intelligence–assisted contouring technology usage | ||
|
Yes | 123 (40.1) |
|
No | 184 (59.9) |
Measurement Model
To test the reliability and validity of this study, we examined common method bias, as all items in our survey were collected from the same questionnaire. We connected all of the items to a common latent factor and used a confirmatory factor analysis to conduct a single-factor test [40]. The results showed that the fit of the single-factor model was poor (χ2377=4294.616; χ2/df=11.392; adjusted goodness of fit index=0.354; comparative fit index=0.628; Tucker–Lewis index=0.599; root mean square error of approximation=0.184), indicating that common method bias was insignificant in this study.
We also conducted some tests to examine the constructs’ reliability, and the results are summarized in Table 3. The Cronbach α and composite reliability values were higher than the threshold of 0.7 [41], which indicated good construct reliability. For convergent validity, all item loadings ranged from 0.633 to 0.980, and good composite reliability (>0.7) was achieved [42].
Table 3.
Variable and item | Item loadings | Mean score of the items | Composite reliability | Convergence validity | Cronbach α | |
PEa | 5.969 | 0.976 | 0.910 | .976 | ||
|
PE1 | 0.946 |
|
|
|
|
|
PE2 | 0.970 |
|
|
|
|
|
PE3 | 0.957 |
|
|
|
|
|
PE4 | 0.942 |
|
|
|
|
EEb | 5.669 | 0.970 | 0.891 | .970 | ||
|
EE1 | 0.921 |
|
|
|
|
|
EE2 | 0.933 |
|
|
|
|
|
EE3 | 0.967 |
|
|
|
|
|
EE4 | 0.954 |
|
|
|
|
SIc | 5.481 | 0.938 | 0.790 | .937 | ||
|
SI1 | 0.909 |
|
|
|
|
|
SI2 | 0.906 |
|
|
|
|
|
SI3 | 0.829 |
|
|
|
|
|
SI4 | 0.909 |
|
|
|
|
PRd | 4.418 | 0.826 | 0.544 | .824 | ||
|
PR1 | 0.633 |
|
|
|
|
|
PR2 | 0.735 |
|
|
|
|
|
PR3 | 0.783 |
|
|
|
|
|
PR4 | 0.788 |
|
|
|
|
RBe | 3.528 | 0.925 | 0.805 | .825 | ||
|
RB1 | 0.865 |
|
|
|
|
|
RB2 | 0.941 |
|
|
|
|
|
RB3 | 0.883 |
|
|
|
|
FCf | 5.446 | 0.929 | 0.766 | .926 | ||
|
FC1 | 0.821 |
|
|
|
|
|
FC2 | 0.886 |
|
|
|
|
|
FC3 | 0.899 |
|
|
|
|
|
FC4 | 0.892 |
|
|
|
|
BIg | 5.484 | 0.980 | 0.942 | .980 | ||
|
BI1 | 0.952 |
|
|
|
|
|
BI2 | 0.980 |
|
|
|
|
|
BI3 | 0.979 |
|
|
|
|
UBh | 1.713 | 0.820 | 0.604 | .800 | ||
|
UB1 | 0.736 |
|
|
|
|
|
UB2 | 0.860 |
|
|
|
|
|
UB3 | 0.729 |
|
|
|
|
aPE: performance expectancy.
bEE: effort expectancy.
cSI: social influence.
dPR: perceived risk.
eRB: resistance bias.
fFC: facilitating conditions.
gBI: behavioral intention.
hUB: use behavior.
Convergent validity was assessed with average variance extracted (AVE) values; all construct AVE values were greater than 0.5 [43] and thus showed good convergent validity. In Table 4, the square roots of the AVE values were higher than the numbers in the off-diagonal direction (correlations between a particular construct in the same column and other constructs in different rows) in the corresponding columns, indicating the good discrimination validity of all constructs. The results of a single-factor analysis of variance showed that the main effect that experience had on behavioral intention was not significant (F3,303=2.587; P=.053).
Table 4.
Constructs | Performance expectancy | Effort expectancy | Social influence | Facilitating conditions | Perceived risk | Resistance bias | Behavioral intention | Use behavior |
Performance expectancy | 0.954 | —b | — | — | — | — | — | — |
Effort expectancy | 0.767 | 0.944 | — | — | — | — | — | — |
Social influence | 0.717 | 0.784 | 0.889 | — | — | — | — | — |
Facilitating conditions | 0.665 | 0.712 | 0.716 | 0.875 | — | — | — | — |
Perceived risk | −0.334 | −0.378 | −0.358 | −0.314 | 0.738 | — | — | — |
Resistance bias | −0.373 | −0.435 | −0.360 | −0.368 | 0.563 | 0.897 | — | — |
Behavioral intention | 0.724 | 0.750 | 0.782 | 0.802 | −0.369 | −0.394 | 0.971 | — |
Use behavior | 0.600 | 0.644 | 0.611 | 0.630 | −0.326 | −0.424 | 0.785 | 0.777 |
aThe items on the diagonal represent the square root of the average variance extracted value; the off-diagonal elements are correlation estimates.
bNot applicable.
The mean scores for behavioral intention (5.484 out of 7.0) and use behavior (1.713 out of 2.0) were high, which showed that the participants had high intentions to use the AI-assisted contouring system. However, even though the total degree of resistance was low, 17.6% (54/307) of the respondents showed a degree of resistance (ranked as an average score of ≥5 for resistance bias). Specifically, 13 out of the 123 (10.6%) participants who had accessed the AI system showed a degree of resistance, and 41 out of the 184 (22.3%) participants who had not accessed the AI system also showed a degree of resistance.
Structural Model Analysis
The fitness of the model was evaluated with the following criteria: (1) the standardized root mean square residual (0.0317) had to be smaller than 0.08, (2) the comparative fit index (0.968) had to be greater than 0.90, and (3) the root mean square error of approximation (0.056) had to also be smaller than 0.08. As shown in Table 5, most of the indicators satisfied the criteria and the combination rule [44], which demonstrated that the hypothesized model was a good fit to the data.
Table 5.
Model fit indices | Criteria | Research model fit values |
Chi-square value | The smaller the better | 692.543 |
Degrees of freedom | The larger the better | 354 |
Normed chi-square value (χ2/df) | 1<χ2/df<3 | 1.956 |
Root mean square error of approximation | <0.08 | 0.056 |
Standardized root mean square residual | <0.08 | 0.0317 |
Comparative fit index | >0.9 | 0.968 |
Goodness of fit index | >0.85 | 0.859 |
Adjusted goodness of fit index | >0.8 | 0.827 |
Figure 2 shows a graphic description of the regression analysis, and Table 6 shows the numerical results of the path coefficient analysis. Standardized factor loadings (ie, β) were the standardized regression coefficients for each proposed path, which indicated the relationships among the variables in this study. P values were the parameters used to determine the results of the hypothesis test [45]. Behavioral intention was positively affected by performance expectancy (β=.155; P=.014), social influence (β=.365; P<.001), and facilitating conditions (β=.459; P<.001), and use behavior was significantly affected by behavioral intention (β=.235; P<.001). Effort expectancy (β=.055; P=.450), perceived risk (β=−.048; P=.348), and resistance bias (β=−.020; P=.634) did not significantly affect behavioral intention, and facilitating conditions (β=.004; P=.845) did not significantly affect use behavior. Behavioral intention also moderated the effects that social influence (P=.001) and facilitating conditions (P<.001) had on use behavior. R2 values were calculated to assess the validity of the research model; 78.1% of the effects of behavioral intention could be explained by perceived risk, effort expectancy, social influence, facilitating conditions, performance expectancy, and resistance bias, and 75.3% of the effects of use behavior could be explained by behavioral intention and facilitating conditions (Figure 2).
Table 6.
Variables and hypotheses | Standardized factor loadings (SE) | Critical ratio (T value) | P value | Results |
Hypothesis 1: Behavioral intention is affected by performance expectancy. | 0.155 (0.063) | 2.464 | .01 | Supported |
Hypothesis 2: Behavioral intention is affected by effort expectancy. | 0.055 (0.073) | 0.755 | .45 | Not supported |
Hypothesis 3: Behavioral intention is affected by social influence. | 0.365 (0.076) | 4.795 | <.001 | Supported |
Hypothesis 4: Use behavior is affected by behavioral intention. | 0.235 (0.024) | 9.778 | <.001 | Supported |
Hypothesis 5A: Use behavior is affected by facilitating conditions. | 0.004 (0.022) | 0.195 | .85 | Not supported |
Hypothesis 5B: Behavioral intention is affected by facilitating conditions. | 0.459 (0.061) | 7.561 | <.001 | Supported |
Hypothesis 6: Behavioral intention is affected by perceived risk. | −0.048 (0.051) | −0.939 | .35 | Not supported |
Hypothesis 7: Behavioral intention is affected by resistance bias. | −0.020 (0.042) | −0.476 | .63 | Not supported |
Discussion
Principal Findings
In this study, we investigated factors that may affect the adoption of an AI-assisted technology for developing treatment plans for cancer radiation therapy patients. The constructs of the research model were developed based on the UTAUT model, and perceived risk and resistance bias were added as potential barriers to new technology adoption. There were 3 principal findings, as follows: (1) the physicians’ overall perceptions of the AI-assisted technology for radiation contouring were high; (2) social influence and facilitating conditions played a much more important role in AI adoption among radiation oncologists than performance expectancy, but effort expectancy was not significantly related to behavioral intention (P=.45); and (3) perceived risk (P=.35) and resistance bias (P=.63) were not significantly related to behavioral intention.
The Physicians’ Overall Perceptions of the AI-Assisted Technology for Radiation Contouring Were High
Participants had high behavioral intentions (score: mean 5.484 out of 7.0) to use the new AI-assisted contouring technology; 90.9% (279/307) of the total sample gave scores of 4 or above, meaning that they were planning to use the system within 1 year. This result nearly matched those of a survey conducted in 2019, in which 84.62% of the participating medical departments planned to try the AI-assisted contouring system; 10.26% of these departments were in the process of purchasing the system, and 46.15% had purchased plans [46].
Participants’ performance expectancy and effort expectancy were high, as they believed that the AI technology could improve their work performance and efficiency. This is similar to the perceptions of Canadian radiation oncologists, who believe that AI could save time and deliver better results to benefit patients [47].
The Relationship Among Performance Expectancy, Social Influence, and Facilitating Conditions Affected Behavioral Intention
Many studies have proven that performance expectancy, effort expectancy, social influence, and facilitating conditions have positive effects on use behavior [8,11,20,48]. For example, in a study of health care robot adoption, these four constructs significantly predicted the usage and uptake of robots [30]. In our study, performance expectancy (P=.01), social influence (P<.001), and facilitating conditions (P<.001) were significantly related to behavioral intention, while effort expectancy did not significantly affect behavioral intention (P=.45). Social influence and facilitating conditions affected behavioral intention much more strongly than performance expectancy. These results indicate that in China, when physicians encounter AI-assisted technology for treatment plans, their opinions are more likely to be driven by people with which they have a close relationship (ie, the items of the social influence construct), such as friends and colleagues, department heads, tutors, superiors, and hospital leaders. This phenomenon reflects the culture of vertical collectivism, which is aligned with the ideology of “utilitarian Guanxi”—a Chinese philosophy that combines the pursuit of profit with objective goals [49]. Further, social propaganda, such as news stories about successfully using AI technology from professionals in the health care industry, likely influences physicians’ perceptions toward using such technology.
Facilitating conditions did not have a direct effect on use behavior. However, this construct was the most important factor that affected behavioral intention. As shown by our results, behavioral intention moderates the effect that facilitating conditions have on use behavior (P<.001), indicating that providing support, such as support from the information technology team in a hospital, and having the knowledge and skills needed for using new technology were necessarily important and influenced radiation oncologists’ use intentions. On the other hand, most researchers that use the UTAUT model agree that effort expectancy has a significant effect on behavioral intention. However, there have been some exceptions. For example, perceived ease of use did not have an effect on attitudes related to health care professionals’ decisions to accept telemedicine technology [50]. In our study, effort expectancy did not affect use intention. In China, physicians in level 3 public hospitals (the highest standard in China) usually have heavy workloads. For example, at Sun Yat-sen University Cancer Center, the Department of Radiation Oncology works on approximately 1200 cases each day. The participants claimed that they did not mind putting in more effort and spending more time on learning about the new technology in order to later reduce their workloads. Determining whether this approach reduces or redistributes workloads is an area worthy of future investigation. Additionally, the AI technology was developed based on participants’ previously used software. As a result, they found it easier to gain fluency with the system.
Perceived Risk and Resistance Bias Were Not Significantly Related to Behavioral Intention
In most research on new technology adoption, perceived risk has negatively affected the public’s behavioral intentions [11,29]. However, in our study, perceived risk did not affect radiation oncologists’ behavioral intentions. Even though they perceived risks (score: mean 4.418 out of 7.0) such as the possibility of being misled with inaccurate contouring and privacy issues, oncologists still had the intention to use the system, as did others (colleagues and superiors). The main reason for this was that it could effectively reduce their heavy workloads. In interviews with some of the participants, physicians explained that they could manually fix the errors in the contouring area because the amount of patient case data has increased and machine learning has made the AI-assisted contouring system more and more accurate; therefore, more time will be saved in the future.
As discussed in a study by Ye et al [10], resistance bias has had a negative impact on the technological acceptance of AI ophthalmic devices. However, different from the general public’s perceptions, the low score for resistance bias (score: mean 3.528 out of 7.0) in our study showed that radiation oncologists were not likely to reject AI-assisted contouring technology, regardless of their unfamiliarity with the technology, their past experiences with the failure of new technology products, or the possibility of AI replacing and displacing jobs.
Our study demonstrated that resistance to using an AI-based contouring system to improve patient outcomes and work efficiency was low. There was little to no apprehension among users to learn about and implement this new technology. In previous studies of technology acceptance, resistance bias was observed among medical staff when the benefits of technology were not clear or immediate [36,51,52]. In past research on health care robot adoption, resistance bias did not have a direct negative effect on medical staff’s behavioral intentions, as the health care robots were convenient to use and provided significantly better results than those provided by traditional methods [30]. In our study, the AI platform was developed by using the previous software that the oncologists used for work. This resulted in immediate benefits for improving contouring results and saving time. Therefore, negative perceptions surrounding new technologies with clear and immediate benefits, particularly fear and apprehension toward using such technologies (regardless of the degree of these perceptions), may not be prevalent in China. On the other hand, the AI contouring system in our study is used as an assistance tool for providing image contouring results as part of the process for developing a radiation therapy plan. In our case, oncologists still have to work on modifying the contouring areas and deciding on the amounts of radiation for each area. Wong et al [47] conducted a study among Canadian radiation oncologists that showed similar results; 77% of the participants agreed that human oversight was necessary for optimizing the planning process and delivering better quality care.
In China, using AI-assisted technology to relieve physicians’ workloads will be a major trend in the future, as such technology tries to fill the gap between the public’s demands and the limited amount of medical resources. Chinese hospitals have the need and motivation to adopt AI to improve efficiency, solve the health care professional shortage problem, and reduce patient waiting times. At Sun Yat-sen University Cancer Center, a breakthrough point for the AI product was the potential research opportunities with the AI development company for improving treatment planning accuracy, as the AI product not only helps oncologists improve their work efficiency but also serves as a valuable tool for scientific research.
Hospitals in China, like those in other Asian countries such as India, South Korea, and Thailand, have also proudly indicated their usage and acceptance of AI technology in their marketing campaigns, which have expounded to patients how much their medical services are improved by such technologies [18]. However, anecdotal reports on AI products in cancer care developed by IBM Corporation, one of the largest AI technology providers in the industry, have shown that the company has had trouble with finding customers in the United States [53] and that technology resistance exists among oncologists who claim that they trust their own judgment over that of software [18]. We suggest that cultural differences may result in different perceptions on AI adoption in Western and Asian countries. Asian physicians may be more likely to accept the idea that AI technology is an assistant that performs certain routine tasks and more likely to reject AI doctors. In this case, the perception is that the AI system will not displace or replace the clinical practitioner; instead, the system will get rid of routine tasks and thus allow for more time to focus on patient care.
Implications for Practice
This study provides 3 suggestions for health care AI developers and hospitals that are planning to introduce AI for clinical decision-making. First, when a hospital introduces a new AI-assisted technology for clinical decision-making in China, the leadership should consider a top-down management process. Hospital leaders and department heads should demonstrate how a new technology works, advocate for the new technology, and encourage subordinate physicians to use the new technology. The hospital could also share AI success stories to influence public opinions on the use of the new technology and highlight the benefits of the system. At Sun Yat-sen University Cancer Center, breakthrough news stories about AI clinical research have been shown not only on social media platforms and local news programs but also at special events such as AI versus human competitions, which have been held at a number of academic conferences, to prove the accuracy of their AI system to both the public and experts. Second, the user interface of the new AI-assisted system should be as consistent as possible with the previously used interface to reduce the time required for users to adapt to the new technology. The hospital information technology center should also work closely with AI technology companies to provide technical support and data security for the use of the new AI platform. Third, per the lesson learned from IBM Corporation’s failure in the United States, AI technology development (ie, progressing from perceptual intelligence to cognitive intelligence) should be a step-by-step process. IBM Corporation’s oncology expert advisor project used natural language processing to provide treatment recommendations. However, the ambitions of such project were, in hindsight, a step too far. Instead of installing an AI doctor, Chinese hospitals first created an AI technology to act as an assistant tool for medical imaging to free physicians from routine work. When a clear application scenario for medical AI is imagined and created, people are more likely to realize the value of such technology.
Limitations and Future Research
This study has some limitations. Our results might be biased due to the age of participants, as 87.6% (269/307) of the participants were aged under 40 years. The majority of the participants (176/307, 57.3%) were also from Sun Yat-sen University Cancer Center, where medical capacity is high. Additionally, there might be elements of sampling bias, as convenience sampling was used. Further research needs to be carried out among medical professionals in other countries for comparisons of AI technology acceptance in health care services in different cultural contexts.
Conclusion
In this study, technology resistance among Chinese radiation oncologists was low and not related to behavioral intention. Negative perceptions of new technologies with clear and immediate benefits, such as fear and apprehension toward using such technologies (regardless of the degree of these perceptions), may not be prevalent in China. Further, not all of the factors in the Venkatesh UTAUT model applied to AI technology adoption among physicians in a Chinese context. Oncologists’ opinions are more likely to be driven by people with which they have a close relationship (the items of the SI construct), such as friends and colleagues, department heads, tutors, superiors, and hospital leaders, and be based on receiving technology support from information technology departments and companies.
Acknowledgments
This study was supported by the National Key Research and Development Program of China (grants 2020YFC1316900 and 2020YFC1316904), the Young Creative Talent Program of Sun Yat-sen University Cancer Center (grant PT21100201), the PhD Start-up Fund of the Natural Science Foundation of the Guangdong Province of China (grant 2018A030310005), the China Postdoctoral Science Foundation (grant 2019M663348), and the National Natural Science Foundation of China (grants 72102238, 71572207, 71832015, and 72072191). The authors would like to thank Ms Wanfei Xie and Ms Le Xia from Sun Yat-sen University Cancer Center for their help with questionnaire collection.
Abbreviations
- AI
artificial intelligence
- AVE
average variance extracted
- TAM
Technology Acceptance Model
- UTAUT
Unified Theory of Acceptance and Use of Technology
Footnotes
Conflicts of Interest: None declared.
References
- 1.Ahuja AS. The impact of artificial intelligence in medicine on the future role of the physician. PeerJ. 2019;7:e7702. doi: 10.7717/peerj.7702. doi: 10.7717/peerj.7702.7702 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.2020 Blue Book of Artificial Intelligence Medical Industry Development. China Academy of Information and Communication. 2020. [2021-07-16]. http://www.caict.ac.cn/kxyj/qwfb/ztbg/202009/t20200908_323708.htm .
- 3.White paper on global development of artificial intelligence. Deloitte China. 2019. [2020-11-16]. https://www2.deloitte.com/content/dam/Deloitte/cn/Documents/technology-media-telecommunications/deloitte-cn-tmt-ai-report-zh-190919.pdf .
- 4.Reddy S, Fox J, Purohit MP. Artificial intelligence-enabled healthcare delivery. J R Soc Med. 2019 Jan;112(1):22–28. doi: 10.1177/0141076818815510. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Musen MA, Shahar Y, Shortliffe EH. Clinical decision-support systems. In: Inhortliffe EH, Cimino JJ, editors. Biomedical Informatics: Computer Applications in Health Care and Biomedicine, 3rd Edition. New York, New York: Springer; 2006. pp. 698–736. [Google Scholar]
- 6.Teguh DN, Levendag PC, Voet PW, Al-Mamgani A, Han X, Wolf TK, Hibbard LS, Nowak P, Akhiat H, Dirkx ML, Heijmen BJ, Hoogeman MS. Clinical validation of atlas-based auto-segmentation of multiple target volumes and normal tissue (swallowing/mastication) structures in the head and neck. Int J Radiat Oncol Biol Phys. 2011 Nov 15;81(4):950–957. doi: 10.1016/j.ijrobp.2010.07.009.S0360-3016(10)00932-6 [DOI] [PubMed] [Google Scholar]
- 7.Chen AM, Chin R, Beron P, Yoshizaki T, Mikaeilian AG, Cao M. Inadequate target volume delineation and local-regional recurrence after intensity-modulated radiotherapy for human papillomavirus-positive oropharynx cancer. Radiother Oncol. 2017 Jun;123(3):412–418. doi: 10.1016/j.radonc.2017.04.015.S0167-8140(17)30156-1 [DOI] [PubMed] [Google Scholar]
- 8.Lin L, Dou Q, Jin YM, Zhou GQ, Tang YQ, Chen WL, Su BA, Liu F, Tao CJ, Jiang N, Li JY, Tang LL, Xie CM, Huang SM, Ma J, Heng PA, Wee JTS, Chua MLK, Chen H, Sun Y. Deep learning for automated contouring of primary tumor volumes by MRI for nasopharyngeal carcinoma. Radiology. 2019 Jun;291(3):677–686. doi: 10.1148/radiol.2019182012. [DOI] [PubMed] [Google Scholar]
- 9.Defnet AM, Davis SS. The robot in general surgery-change is the only constant. JAMA Surg. 2021 Jan 01;156(1):30. doi: 10.1001/jamasurg.2020.4590.2771574 [DOI] [PubMed] [Google Scholar]
- 10.Ye T, Xue J, He M, Gu J, Lin H, Xu B, Cheng Y. Psychosocial factors affecting artificial intelligence adoption in health care in China: Cross-sectional study. J Med Internet Res. 2019 Oct 17;21(10):e14316. doi: 10.2196/14316. https://www.jmir.org/2019/10/e14316/ v21i10e14316 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Ataka K, Chen D, Yamaguchi T, Yamamoto I. Living in harmony with artificial intelligence: Will AI replace your work? Journal of Information Processing and Management. 2018;60(12):865–881. doi: 10.1241/johokanri.60.865. [DOI] [Google Scholar]
- 12.Goldhahn J, Rampton V, Spinas GA. Could artificial intelligence make doctors obsolete? BMJ. 2018 Nov 07;363:k4563. doi: 10.1136/bmj.k4563. [DOI] [PubMed] [Google Scholar]
- 13.Lapointe L, Rivard S. Getting physicians to accept new information technology: insights from case studies. CMAJ. 2006 May 23;174(11):1573–1578. doi: 10.1503/cmaj.050281. http://www.cmaj.ca/cgi/pmidlookup?view=long&pmid=16717265 .174/11/1573 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Bhattacherjee A, Hikmet N, Menachemi N, Kayhan VO, Brooks RG. The differential performance effects of healthcare information technology adoption. Information Systems Management. 2006 Dec 22;24(1):5–14. doi: 10.1080/10580530601036778. [DOI] [Google Scholar]
- 15.Bhattacherjee A, Hikmet N. Physicians' resistance toward healthcare information technology: a theoretical model and empirical test. Eur J Inf Syst. 2017 Dec 19;16(6):725–737. doi: 10.1057/palgrave.ejis.3000717. [DOI] [Google Scholar]
- 16.Hsieh PJ. Physicians' acceptance of electronic medical records exchange: an extension of the decomposed TPB model with institutional trust and perceived risk. Int J Med Inform. 2015 Jan;84(1):1–14. doi: 10.1016/j.ijmedinf.2014.08.008.S1386-5056(14)00162-2 [DOI] [PubMed] [Google Scholar]
- 17.Noorbakhsh-Sabet N, Zand R, Zhang Y, Abedi V. Artificial intelligence transforms the future of health care. Am J Med. 2019 Jul;132(7):795–801. doi: 10.1016/j.amjmed.2019.01.017. http://europepmc.org/abstract/MED/30710543 .S0002-9343(19)30120-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Strickland E. How IBM Watson overpromised and underdelivered on AI health care. IEEE Spectrum. 2019. Apr 02, [2020-12-16]. https://spectrum.ieee.org/biomedical/diagnostics/how-ibm-watson-overpromised-and-underdelivered-on-ai-health-care .
- 19.Cooper L, Lombardo C. IBM explores sale of IBM Watson Health. Morningstar. 2021. Feb 18, [2021-07-01]. https://www.morningstar.com/news/dow-jones/2021021816650/ibm-explores-sale-of-ibm-watson-health-update .
- 20.Davis FD. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. 1989 Sep;13(3):319–340. doi: 10.2307/249008. [DOI] [Google Scholar]
- 21.Venkatesh V, Davis FD. A theoretical extension of the technology acceptance model: Four longitudinal field studies. Manage Sci. 2000 Feb;46(2):186–204. doi: 10.1287/mnsc.46.2.186.11926. [DOI] [Google Scholar]
- 22.Venkatesh V, Bala H. Technology acceptance model 3 and a research agenda on interventions. Decis Sci. 2008 May 09;39(2):273–315. doi: 10.1111/j.1540-5915.2008.00192.x. [DOI] [Google Scholar]
- 23.Venkatesh V, Morris M, Davis G, Davis F. User acceptance of information technology: Toward a unified view. MIS Q. 2003;27(3):425–478. doi: 10.2307/30036540. [DOI] [Google Scholar]
- 24.Venkatesh V, Thong JY, Xu X. Consumer acceptance and use of information technology: Extending the unified theory of acceptance and use of technology. MIS Q. 2012;36(1):157–178. doi: 10.2307/41410412. [DOI] [Google Scholar]
- 25.Kijsanayotin B, Pannarunothai S, Speedie SM. Factors influencing health information technology adoption in Thailand's community health centers: applying the UTAUT model. Int J Med Inform. 2009 Jun;78(6):404–416. doi: 10.1016/j.ijmedinf.2008.12.005.S1386-5056(08)00210-4 [DOI] [PubMed] [Google Scholar]
- 26.Shiferaw KB, Mehari EA. Modeling predictors of acceptance and use of electronic medical record system in a resource limited setting: Using modified UTAUT model. Inform Med Unlocked. 2019;17:100182. doi: 10.1016/j.imu.2019.100182. https://www.sciencedirect.com/science/article/pii/S235291481930067X . [DOI] [Google Scholar]
- 27.Bawack RE, Kala Kamdjoug JR. Adequacy of UTAUT in clinician adoption of health information systems in developing countries: The case of Cameroon. Int J Med Inform. 2018 Jan;109:15–22. doi: 10.1016/j.ijmedinf.2017.10.016.S1386-5056(17)30384-2 [DOI] [PubMed] [Google Scholar]
- 28.Jang SH, Kim RH, Lee CW. Effect of u-healthcare service quality on usage intention in a healthcare service. Technol Forecast Soc Change. 2016 Dec;113:396–403. doi: 10.1016/j.techfore.2016.07.030. [DOI] [Google Scholar]
- 29.Martins C, Oliveira T, Popovič A. Understanding the internet banking adoption: A unified theory of acceptance and use of technology and perceived risk application. Int J Inf Manage. 2014 Feb;34(1):1–13. doi: 10.1016/j.ijinfomgt.2013.06.002. [DOI] [Google Scholar]
- 30.Vichitkraivin P, Naenna T. Factors of healthcare robot adoption by medical staff in Thai government hospitals. Health Technol (Berl) 2020 Nov 06;11(1):139–151. doi: 10.1007/s12553-020-00489-4. [DOI] [Google Scholar]
- 31.Berg M, Aarts J, van der Lei J. ICT in health care: sociotechnical approaches. Methods Inf Med. 2003;42(4):297–301.03040297 [PubMed] [Google Scholar]
- 32.Bauer RA. Consumer behavior as risk taking. Dynamic Marketing for a Changing World; The 43rd National Conference of the American Marketing Association; June 15-17, 1960; Chicago, Illinois. 1960. pp. 389–398. [Google Scholar]
- 33.Featherman MS, Pavlou PA. Predicting e-services adoption: a perceived risk facets perspective. Int J Hum Comput Stud. 2003 Oct;59(4):451–474. doi: 10.1016/S1071-5819(03)00111-3. [DOI] [Google Scholar]
- 34.Dou K, Yu P, Deng N, Liu F, Guan Y, Li Z, Ji Y, Du N, Lu X, Duan H. Patients' acceptance of smartphone health technology for chronic disease management: A theoretical model and empirical test. JMIR Mhealth Uhealth. 2017 Dec 06;5(12):e177. doi: 10.2196/mhealth.7886. https://mhealth.jmir.org/2017/12/e177/ v5i12e177 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Guo X, Sun Y, Wang N, Peng Z, Yan Z. The dark side of elderly acceptance of preventive mobile health services in China. Electronic Markets. 2012 Dec 11;23(1):49–61. doi: 10.1007/s12525-012-0112-4. [DOI] [Google Scholar]
- 36.Deng Z, Mo X, Liu S. Comparison of the middle-aged and older users' adoption of mobile health services in China. Int J Med Inform. 2014 Mar;83(3):210–224. doi: 10.1016/j.ijmedinf.2013.12.002.S1386-5056(13)00249-9 [DOI] [PubMed] [Google Scholar]
- 37.Raza MH, Adenola AF, Nafarieh A, Robertson W. The slow adoption of cloud computing and IT workforce. Procedia Comput Sci. 2015;52:1114–1119. doi: 10.1016/j.procs.2015.05.128. [DOI] [Google Scholar]
- 38.Andrews L, Gajanayake R, Sahama T. The Australian general public's perceptions of having a personally controlled electronic health record (PCEHR) Int J Med Inform. 2014 Dec;83(12):889–900. doi: 10.1016/j.ijmedinf.2014.08.002.S1386-5056(14)00145-2 [DOI] [PubMed] [Google Scholar]
- 39.Anderson JC, Gerbing DW. Structural equation modeling in practice: A review and recommended two-step approach. Psychol Bull. 1988 May;103(3):411–423. doi: 10.1037/0033-2909.103.3.411. [DOI] [Google Scholar]
- 40.Malhotra NK, Kim SS, Patil A. Common method variance in IS research: A comparison of alternative approaches and a reanalysis of past research. Manage Sci. 2006 Dec;52(12):1865–1883. doi: 10.1287/mnsc.1060.0597. [DOI] [Google Scholar]
- 41.Rivard S, Huff SL. Factors of success for end-user computing. Commun ACM. 1988 May;31(5):552–561. doi: 10.1145/42411.42418. [DOI] [Google Scholar]
- 42.Hair JF, Ringle CM, Sarstedt M. PLS-SEM: Indeed a silver bullet. Journal of Marketing Theory and Practice. 2014 Dec 08;19(2):139–152. doi: 10.2753/MTP1069-6679190202. [DOI] [Google Scholar]
- 43.Fornell C, Larcker DF. Structural equation models with unobservable variables and measurement error: Algebra and statistics. J Mark Res. 1981 Aug;18(3):382–388. doi: 10.2307/3150980. [DOI] [Google Scholar]
- 44.Hu L, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Struct Equ Modeling. 1999 Jan;6(1):1–55. doi: 10.1080/10705519909540118. [DOI] [Google Scholar]
- 45.Wasserstein RL, Lazar NA. The ASA statement on p-values: Context, process, and purpose. Am Stat. 2016 Jun 09;70(2):129–133. doi: 10.1080/00031305.2016.1154108. [DOI] [Google Scholar]
- 46.2019 China Radiotherapy AI Contouring Product Usage Survey Report. Radiotherapy Front. 2019. [2020-12-23]. https://mp.weixin.qq.com/s/6N9tMen7768OxpmMmzJRvg .
- 47.Wong K, Gallant F, Szumacher E. Perceptions of Canadian radiation oncologists, radiation physicists, radiation therapists and radiation trainees about the impact of artificial intelligence in radiation oncology - national survey. J Med Imaging Radiat Sci. 2021 Mar;52(1):44–48. doi: 10.1016/j.jmir.2020.11.013.S1939-8654(20)30381-7 [DOI] [PubMed] [Google Scholar]
- 48.Chang I, Hwang H, Hung W, Li Y. Physicians’ acceptance of pharmacokinetics-based clinical decision support systems. Expert Syst Appl. 2007 Aug;33(2):296–303. doi: 10.1016/j.eswa.2006.05.001. [DOI] [Google Scholar]
- 49.Du J, Vantilborgh T. Cultural differences in the content of employees' psychological contract: A qualitative study comparing Belgium and China. Psychol Belg. 2020 Jun 01;60(1):132–151. doi: 10.5334/pb.498. http://europepmc.org/abstract/MED/32523710 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Chau PY, Hu PJ. Investigating healthcare professionals’ decisions to accept telemedicine technology: an empirical test of competing theories. Information & Management. 2002 Jan;39(4):297–311. doi: 10.1016/S0378-7206(01)00098-2. [DOI] [Google Scholar]
- 51.Panteleimonitis S, Parvaiz A. Robotic colorectal surgery. Surgery (Oxf) 2017 Mar;35(3):151–156. doi: 10.1016/j.mpsur.2016.12.004. [DOI] [Google Scholar]
- 52.Brewster L, Mountain G, Wessels B, Kelly C, Hawley M. Factors affecting front line staff acceptance of telehealth technologies: a mixed-method systematic review. J Adv Nurs. 2014 Jan;70(1):21–33. doi: 10.1111/jan.12196. [DOI] [PubMed] [Google Scholar]
- 53.Ross C, Swetlitz I. IBM pitched its Watson supercomputer as a revolution in cancer care. It’s nowhere close. STAT. 2017. Sep 05, [2020-12-23]. https://www.statnews.com/2017/09/05/watson-ibm-cancer/