Skip to main content
BMC Medical Informatics and Decision Making logoLink to BMC Medical Informatics and Decision Making
. 2013 Jan 5;13:2. doi: 10.1186/1472-6947-13-2

User needs elicitation via analytic hierarchy process (AHP). A case study on a Computed Tomography (CT) scanner

Leandro Pecchia 1,, Jennifer L Martin 1, Angela Ragozzino 2, Carmela Vanzanella 3, Arturo Scognamiglio 4, Luciano Mirarchi 5, Stephen P Morgan 1
PMCID: PMC3545827  PMID: 23289426

Abstract

Background

The rigorous elicitation of user needs is a crucial step for both medical device design and purchasing. However, user needs elicitation is often based on qualitative methods whose findings can be difficult to integrate into medical decision-making. This paper describes the application of AHP to elicit user needs for a new CT scanner for use in a public hospital.

Methods

AHP was used to design a hierarchy of 12 needs for a new CT scanner, grouped into 4 homogenous categories, and to prepare a paper questionnaire to investigate the relative priorities of these. The questionnaire was completed by 5 senior clinicians working in a variety of clinical specialisations and departments in the same Italian public hospital.

Results

Although safety and performance were considered the most important issues, user needs changed according to clinical scenario. For elective surgery, the five most important needs were: spatial resolution, processing software, radiation dose, patient monitoring, and contrast medium. For emergency, the top five most important needs were: patient monitoring, radiation dose, contrast medium control, speed run, spatial resolution.

Conclusions

AHP effectively supported user need elicitation, helping to develop an analytic and intelligible framework of decision-making. User needs varied according to working scenario (elective versus emergency medicine) more than clinical specialization. This method should be considered by practitioners involved in decisions about new medical technology, whether that be during device design or before deciding whether to allocate budgets for new medical devices according to clinical functions or according to hospital department.

Keywords: User needs elicitation, Analytic hierarchy process, AHP, Medical decision-making, Medical device

Background

To provide high quality care for patients, the healthcare industry is dependent upon the provision of complex and expensive medical devices. It is widely accepted that if devices are to be used effectively they must meet the requirements of their users [1], however, capturing user requirements for healthcare technology is extremely complex. Although clinical effectiveness and safety are the primary concerns in medicine, many other aspects must also be considered including training needs, storage, labelling, servicing and cleaning [2]. Moreover, for the same medical device, the concepts of effectiveness and safety may change according to the specific clinical problem, medical specialization and patient condition.

The topic of user requirements of medical devices is of interest to a wide variety of individuals and organisations that are required to make decisions on the development, purchasing and prescription of these products. However, research has shown that collecting and considering this information is a challenging undertaking; a lack of time and resources may preclude rigorous work into requirements [3], as can a lack of knowledge of appropriate methods for data collection and analysis [4]. This can result in the collection of data that are incomplete, difficult to interpret or that fail to address the questions of interest [5].

Finally, and most fundamentally, the complex nature of medical device user requirements means that for any one medical device there are likely to be a large number of possible users, potentially including both professional and lay users, all with differing specialities, skills and abilities. Even within seemingly homogeneous user groups, individuals will have received different training and will vary in their working patterns, attitudes and preferences. In addition, how a device is used will vary considerably, according to the particular clinical procedure being performed and the physical and organisational context in which it is used [2]. This information must not only be collected and considered, but differences and conflicts between users must also be balanced. This is a critical issue for the developers of medical technology but also for healthcare providers when making purchasing decisions. It is a particular issue for publically funded healthcare providers who must demonstrate that the purchasing decisions about high-cost equipment are transparent and are be based on the best possible evidence available at the time.

The use of scientific quantitative methods to support decision making is considered necessary in healthcare organizations, where the personnel are committed to follow only the best available evidence according to well-designed trials [6], meta-analyses [7] or network meta-analyses [8]. Nonetheless, despite the hierarchy of evidence, the complexities of medical device decision-making require a spectrum of qualitative and quantitative information [9]. At the start of a user need elicitation problem, a wide-ranging and open-ended study should be conducted to collect data about the needs and priorities of healthcare professionals [10]. This type of information is critical to developing a broad understanding of the range of user requirements. In medical decision-making, qualitative methods have a crucial role in examining evidence from previous studies [9,11] and appraising this according to different contexts of use. It has been suggested that improving the methods used in qualitative studies will legitimise this type of data and increase its use in healthcare decision-making [12] as advocated by Kaplan [13], who concluded: “a plea is made for incorporating qualitative/interpretive/subjectivist methods, without prejudice to other approaches”. Furthermore, evidence-based care advocates that medical decisions are made with reference to the best available research evidence [14].

However, the nature of qualitative research can limit its use in scientific decision-making tasks such as user needs requirements elicitation for medical devices. The influence that the researcher plays in designing and interpreting studies has resulted in qualitative methods being viewed with scepticism by the medical community [15]. In addition, researchers have encountered problems when attempting to use qualitative data in the analytic and scientific decision-making processes that are a fundamental part of healthcare research [16]. For example, how can open-ended interview data collected from a number of caregivers with a range of opinions be used to make decisions on the design of a new medical device in a transparent and rigorous way [5]. There is need therefore for new approaches that allow the breadth and depth of the topics under investigation to be captured, yet also allow these to be quantified and prioritised, and for the process to be as transparent as possible. This is not only important for the decision makers but also for the healthcare staff; research has shown that successful adoption of new healthcare technology is dependent upon joint ownership of the decisions made during the development process [17]. Moreover, the decision outcome should be easy to understand, as intelligibility is strongly appreciated in medical domain decision-making [18-20], especially in the public sector. Finally, although not the primary aim of this study, the use of AHP clearly has implications for device manufacturers and future technology strategy in this area. In fact, medical device companies have also demonstrated an interest in scientific methods to elicit user needs, to enable them to respond to clinical demand and to enter new markets by adapting their products to the requirements of different medical specializations [21].

The Analytic Hierarchy Process (AHP) is a multi-dimensional, multi-level and multifactorial decision-making method based on the idea that it is possible to prioritize elements by: grouping them into meaningful categories and sub-categories; performing pairwise comparisons; defining a coherent framework of quantitative and qualitative knowledge; measuring intangible domains. This hierarchical approach allows the construction of a consistent framework for step-by-step decision-making, breaking a complex problem into many small less-complex ones that decision-makers can more easily deal with. This paradigm, known as divide et impera[22] (divide and rule) and widely investigated in medicine [23,24], has been demonstrated to be effective in healthcare decision-making [25].

The AHP is effective for quantifying qualitative knowledge as it allows intangible dimensions such as subjective preferences and comfort to be measured. This is important in medical decision-making as these factors [26], which are normally examined with qualitative research, cannot be measured directly using an absolute scale [27]. The AHP is particularly effective for quantifying experts’ opinions [28] that are based on personal experience and knowledge to design a consistent decision framework. This is a crucial point in any medical context [13], where not all of the relevant information is objective or quantitative. A number of researchers have highlighted the benefits of using AHP to explore user needs in healthcare [29,30], and in particular for including patient opinions in health technology assessment [31,32], choosing treatments [33], and improving patient centred healthcare [34,35]. Other methods that have attempted to elicit and quantify user needs in healthcare are conjoint analysis (CA) [36] , discrete choice experiments [37] and best-worst scaling [38]. A growing number of articles have focused on comparing AHP with these methods, and in particular with CA. According to Scholl et al. [39], AHP has proven to be more suitable than CA for complex decisions involving many factors. Mulye [40] suggested that AHP is more effective than CA when more than 6 attributes have to be prioritized. Ijzerman et al. [41] concluded that AHP, when compared with CA, resulted in more flexible, easier to implement and shorter questionnaires, although it may generate some inconsistences and other methods may have a more holistic approach. In another study, Ijzerman et al. [42], concluded that AHP lead to the overestimation of some alternatives although the differences found between AHP and CA, were mainly ascribed to the labelling of the attributes and the elicitation of performance judgments.

In our elicitation of user needs, we used AHP rather than the methods mentioned above because this method has been applied to medical decision-making [43] at the hospital level for budget allocation [44] and medical device purchasing [45]. It has been shown to be useful for a range of healthcare related decisions and for individuals from a range of backgrounds. As such, this method has the potential to be effective for the different organisations and individuals that are interested in eliciting user requirements, for example: developers wishing to improve device design, hospital managers who must allocate budgets and clinical engineers that are required to select devices. In addition to assisting each of these isolated tasks, a method that could be shown to be usable by all these groups could also improve communication between them, which is also essential in healthcare decision-making. AHP is normally used within a group decision-making process and requires that the decision-makers meet to compare and discuss their weights and decisions as a means to develop a consensus on group weights and achieve a group decision. However, this was not the purpose of this study, which aimed instead to explore the differences between the needs of clinicians with different specializations and different clinical settings. In summary, the adoption of a common method to elicit and prioritise user requirements could facilitate a wide range of decisions related to the design, selection and purchasing of medical devices.

In this study, we focus on clinical user needs related to the use of a multi-slice Computer Tomography (CT) scanner in a medium size city hospital. The multi-slice CT scanner refers to a special CT system equipped with a multiple-row detector array to collect simultaneously data at different slice locations. The multi-slice CT scanner has the capability of rapidly scanning a large longitudinal volume with high resolution. There are two modes for a CT scan: step-and-shoot CT or helical (or spiral) CT [46]. In recent years, developments in CT technology have provided increasing temporal and better spatial resolution. Scan times are much shorter and slice thickness much thinner with increasing rotation speed and increasing number of active detector-rows, from 4 and 16 detector rows to 64-detector CT scanners [47]. The different features of this device may significantly affect its costs. For instance, to equip this device with a system for continuous patient monitoring during the examination may be expensive. In addition, the technical performance of the device may strongly vary, affecting the final cost. It is therefore of paramount importance to elicit user needs before the purchasing decision is made to ensure that the right device is chosen and not one with unnecessary and costly features.

In particular, we focus on the application of AHP to identify the differences between the needs of clinical users, stratifying them according to specialization and intervention (elective versus emergency). We describe how the AHP method was adapted to improve its effectiveness for application in healthcare contexts [21,48], while a more general description of the AHP can be found elsewhere [49].

Methods

Ethical considerations

Before beginning the study the protocol was discussed with the hospital ethical committee. As this was an interview study with clinical staff and without patient involvement, no formal approval by an ethics committee was required. A participant information sheet was presented and discussed with participants before their involvement.

Hierarchy definition

A focus group identified a total of 12 different clinical needs that must be satisfied by a CT-scanner. This focus group involved 4 medical doctors in charge of the units, of which 2 are co-authors of this paper (AR and AS), 3 biomedical engineers with extensive experience of the design, assessment and management of medical devices, of which 2 are co-authors of this paper (LP and LM) and 1 clinical engineer of the hospital. This group identified 12 needs, based on their personal experience and the pertinent scientific literature, and organized them into meaningful categories. LP acted as the facilitator and, based on his experience of AHP, designed the hierarchy, which was then reviewed with the other participants to check that it was accurate and comprehensive.

The 12 needs were organized into four categories and a tree was designed in which each node represented a category, and each leaf represented a need (Figure  1).

Figure 1.

Figure 1

Tree of needs.

Questionnaires

Questionnaires were designed to enable each respondent to compare the relative importance of each need with all of the other needs within the same category. The layout of the questionnaire is illustrated in Figure  2.

Figure 2.

Figure 2

Questionnaire layout.

For each pair of needs (ij), responders were asked the following question: “in the selection of a new CT scanner, according to your experience, how important do you consider the element i compared to the element j?”. Responders answered by choosing one of the following judgments: much less, less, equally, more, or much more important. In accordance with the Saaty natural scale [50], an integer numerical value was given to each judgment: 1 if equally, 3 if more important, and 5 if much more important. The reciprocal values were given to the remaining judgments: 1/3 if less important, 1/5 if much less important. In-between numbers were used for in-between judgments. Although several scales have been proposed for this process [51-53], in this study an adaptation of the Saaty natural scale was used as it is easier to understand for responders who are not skilled in complex mathematics or with the AHP method. In this study, we used a three-point scale and not a nine-point scale as previous studies [27][54], involving approximately 200 responders unskilled in the use of AHP have shown that:

1. Although having a 9-point scale, most responders did not use more than 3 judgments (equal, more, much more) when comparing up to 4 elements.

2. Lay users reported confused when using a more complex scale.

Other studies have utilized a reduced scale (see supplementary material of [42]), although not clearly stated in the method section of the paper. After normalizing the eigenvectors by using the distributive mode [49], the results achieved with a five-point scale are equivalent to those achieved using the nine-point fundamental scale. These results were presented in four articles at recent International Symposia on AHP (ISAHP) [55,56] and [26].

The process was then repeated, designing similar questionnaires to elicit the relative importance of each category of needs. The questionnaire was designed to minimize possible responder bias. As responders writing from left to right and top-down can be more likely to judge the elements on the top-left as more important than those on the bottom right, each element was presented the same number of times on the left and the right, at the top and at the bottom of the questionnaire Moreover, the sequence of comparisons (A with B, B with C and C with A) was adapted to minimize intransitive judgments [54].

Judgment matrix

For each category of needs, a judgment matrix Anxn was designed, where “n” is the number of needs in this category. According to Saaty theory [50], each matrix had the following properties:

1. The generic element (aij) referred to the ratio between the relative importance of the need “i” (Ni) and “j” (Nj);

2. The element aji was the reciprocal of aij, assuming the reciprocity of judgment (if Ni was 3 times more important than Nj, then Nj should be 1/3 of Ni);

3. The element aii was equal to 1 (Ni is equal in importance to itself);

4. The matrix A was assumed to be a transitive matrix, which means that “∀ i, j, k (1; n), aij = aik * akj” by definition of aij (see Equation 1).

aij=NiNj=NiNk*NkNj=aik*akj (1)

This last property is called the transitivity property and reflects the idea that if “i” was considered twice as important as j (Ni= aij * Nj), and “j” was considered three times more important than “k” (Nj= ajk * Nk), then “i” should be judged six times (two times three) more important than “k” (Ni = aik * Nk, with aik=aij* ajk).

Local weights: the relative importance of needs within each category

It has been proved [50] that, if a matrix A satisfies the properties described in section 2.4 then each column is proportional to the others and only one real eigenvalue (λ) exists, which is equal to “n”. The eigenvector associated with this eigenvalue is again proportional to each column, and represents the relative importance of each need compared to each of the other needs in the same category. The relative importance (weight) of a need i within the category m will be further recalled as LWim or local weight.

In cases where the judgments are not fully consistent, the columns of the matrix are not proportional to one another. In addition, the matrix has more eigenvectors and none are proportional to all the columns. In this case, the main eigenvector, which is the one corresponding to the largest eigenvalue (λmax), is chosen. Its normalized components represent the relative importance of each need.

Consistency estimation

If the transitivity property is not respected, an inconsistency will be generated. This inconsistency was estimated by posing some redundant questions. Considering three needs (i, j, and k) the respondent was asked to perform the pair comparisons ij and jk, and then the redundant comparison ik. The answer to the redundant question was compared with the one deduced from the first two, assuming the transitivity of judgment. The difference between the real answer and the transitive one represents the degree of inconsistency. The global effect of this inconsistency was estimated by measuring the difference between the major eigenvalue λmax and “n”. The error is zero when the framework is completely consistent. Inconsistency is, in the majority of cases, due to loss of interest or distraction. If inconsistency occurs, the responders are required to answer the questionnaire again. Some inconsistency between responses is expected; using a scale of natural numbers will cause some systemic inconsistency because not all the ratios can be represented and because of the limited upper value (e.g. 3*2 gives 6, but the maximum value in the scale is 5). For this reason, an error less than a certain threshold was accepted in accordance with the literature [57]. An error over this threshold should be considered too high for reliable decisions.

At each node, the responders’ consistence was estimated measuring the difference of the eigenvalue λmax from “n” (number of elements in the node), normalized to “n”. This is defined as the consistency index (CI) [57], and is zero when the framework is completely consistent (λmax=n). According to literature, the CI is divided by the Random Consistency Index (R.I.), which is a tabled [57] value changing for n from 1 to 9. This ratio is called Consistency Ratio (CR=CI/CR) and a threshold of CR≤ 0.1 is generally considered appropriate, although some authors have proved that it is possible to increase this threshold to 0.2 when the hierarchy is complex and it is not practical for the responders to discuss the questionnaire results [26,54].

Category importance per responder

By applying the same algorithm to the categories it was possible to evaluate their relative importance. The relative importance of a category m will be further recalled as category importance (weight) or Categorical Weight (CWm).

Global-importance of each need per responder

Finally, the relative importance of a need i compared to all the others (not only those in the same category) is defined as global-importance (Global-Weight) of the need i (GWi). GWs are calculated by multiplying the local (within category) importance of the need by the importance of the root element (category) into the Hierarchy. For instance the global-weight of the need i, which is in the category m, was calculated as the product of the local importance of the need (LWik) and the importance of its category m (CWm) (Equation 2).

GWi=LWik*CWk (2)

Correlations among responders’ preferences

The goal of this study was to explore the differences between user needs for a CT scanner, stratifying clinicians according to specialization and type of intervention (elective versus emergency), and not to find consensus between them. Finding consensus usually requires that the group of responders meet to compare and discuss their weights and to agree a group decision. Nonetheless, this study did investigate the correlations between the responses to understand whether needs were more homogeneous according to clinical specialization (i.e. neurologists versus ear surgeon) or according to the type of intervention (elective versus emergency). This is an important issue both for device design and purchasing.

Several methods have been proposed to measure consensus [58], but as stated, this study does not aim to obtain a consensus, but rather to measure correlations to investigate differences in the needs of different users. Thus, the Spearman rank correlation (ρ or RHO) was calculated, as this measure is widely used for AHP-based studies [54,59]. This correlation measures mathematically if two sets of elements are ranked in the same order [39]. Large values of RHO show well-matched rankings (1, identical ranking) of prioritized elements. To verify the significance of ρ, the p-value was used to test the hypothesis that two responders’ prioritizations are meaningfully correlated. A value of p less than 0.05 was considered significant, according to existing literature [54]. Thus, the homogeneity of correlations was tested by calculating the matrix of p-values for testing the hypothesis of no correlation against the alternative that there is a nonzero correlation. Each element of this matrix is the p-value for the corresponding element of RHO. If the p-value (i, j) is less than 0.05, then the correlation RHO (i, j) is significantly different from zero, which in this study meant that responder-i and responder-j prioritized the need in the same order.

User feedback

Finally, to fully understand the reasons behind the needs prioritization, the results obtained were discussed with the responders, other domain experts (clinicians working in similar scenarios to the responders) and the Medical Director of the Trust. Some open questions were also posed to obtain feedback on the method.

Responders

Five clinicians (age 54±5 years, 40% males), each with more than 20 years of experience, working in the same medium-sized public hospital, were the final responders in the study and completed the questionnaires. None of these clinicians was one of the authors of this paper. All had experience of different clinical environments, but each was asked to answer in relation to the unit in which they were working at the time of the study, which were: radiology unit, emergency unit, minimally invasive ear surgery unit, neurology unit. The surgeon from the ear surgery unit was mainly responsible for child ear cochlear implants, which is an elective surgery. Two surgeons answered from the neurology units: one was in charge of emergency neurological surgeries and the other of the elective neurologic surgeries.

Results

The relative importance for each category of needs is reported in Table  1.

Table 1.

Categorical local weights (CR≤0.1)

  radiology ear surgery neurology emergency neurology emergency
PERFORMANCE
0.22
0.44
0.39
0.19
0.18
SAFETY
0.48
0.34
0.29
0.57
0.60
USABILITY
0.19
0.14
0.22
0.17
0.13
TECHNICAL ISSUES 0.11 0.08 0.10 0.08 0.09

The global and local weights of each need are reported in Table  2.

Table 2.

Local and global weight of needs (CR≤0.1)

  radiology ear surgery neurology emergency neurology emergency
 
GW
(LW)
GW
(LW)
GW
(LW)
GW
(LW)
GW
(LW)
PERFORMANCE
 
 
 
 
 
 
 
 
 
 
  Spatial Resolution
.07
(.32)
.28
(.64)
.19
(.48)
.08
(.43)
.05
(.30)
  Speed Run
.10
(.46)
.05
(.11)
.04
(.11)
.08
(.43)
.09
(.52)
  Processing software
.05
(.22)
.11
(.26)
.16
(.41)
.03
(.14)
.03
(.18)
SAFETY
 
 
 
 
 
 
 
 
 
 
  Patient radiation dose
.16
(.33)
.11
(.33)
.10
(.33)
.08
(.14)
.11
(.18)
  Patient Monitoring
.16
(.33)
.11
(.33)
.10
(.33)
.33
(.58)
.40
(.66)
  Contrast medium control
.16
(.33)
.11
(.33)
.10
(.33)
.16
(.28)
.10
(.16)
USABILITY
 
 
 
 
 
 
 
 
 
 
  Personnel Education
.13
(.69)
.07
(.52)
.07
(.32)
.08
(.48)
.04
(.33)
  User-friendly GUI
.04
(.23)
.03
(.18)
.05
(.22)
.02
(.11)
.04
(.33)
  Interoperability
.02
(.08)
.04
(.30)
.10
(.46)
.07
(.41)
.04
(.33)
TECHNICAL ISSUES
 
 
 
 
 
 
 
 
 
 
  Technical Assistance
.04
(.33)
.03
(.33)
.02
(.22)
.05
(.66)
.04
(.46)
  Maintenance
.04
(.33)
.03
(.33)
.03
(.32)
.01
(.16)
.03
(.32)
  Data Storing .04 (.33) .03 (.33) .05 (.46) .01 (.18) .02 (.22)

Table  3 and Table  4 show the relationship between the responders’ prioritization via Spearman rank correlation, according to respectively per category weight and per needs’ global weight.

Table 3.

Spearman correlation (ρ) and p-value (p) among responders per categories prioritization

 
Radiology
ear surgery
Neurology
Emergency neurology
Emergency
  ρ(p) ρ(p) ρ(p) ρ(p) ρ(p)
radiology
1
-
-
1(0.040)
1(0.040)
ear surgery
 
1
1(0.040)
-
-
neurology
 
 
1
-
-
Emergency neurology
 
 
 
1
1(0.040)
emergency         1

Table 4.

Spearman correlation (ρ) and p-value (p) among responders per needs’ GW

  Radiology ear surgery Neurology Emergency neurology Emergency
 
ρ(p)
ρ(p)
ρ(p)
ρ(p)
ρ(p)
radiology
1
0.73(0.004)
-
0.82(0.001)
0.77(0.002)
ear surgery
 
1
0.86(0.000)
0.74(0.003)
0.61(0.018)
neurology
 
 
1
-
-
Emergency neurology
 
 
 
1
0.90(0.000)
emergency         1

All responders achieved the required threshold for coherence (CR≤0.1), as detailed in Table  5.

Table 5.

Consistency ratio (CR) per responder per questionnaire

  radiology ear surgery neurology emergency neurology emergency
Questionnaire 1: PERFORMANCE
0.00
0.04
0.03
0.00
0.01
Questionnaire 2: SAFETY
0.00
0.00
0.00
0.01
0.03
Questionnaire 3: USABILITY
0.10
0.01
0.00
0.03
0.00
Questionnaire 4: TECHNICAL ISSUES
0.00
0.00
0.00
0.03
0.00
Questionnaire 5: CATEGORIES 0.01 0.07 0.06 0.04 0.02

Discussion

In this paper, we presented the results of a study on the application of AHP to elicit clinical user needs. As a case study, we focused on user needs related to the use of a CT scanner in a medium size hospital.

For elective surgery (ear and neurology), technical performance was considered the most important category of needs, while in emergency departments the safety of the patient was the dominant need. Patient safety was considered at least the second most important category by all the clinicians. All the responders considered technical issues the least important category. The results in Table  1 show that the relative importance of each category of needs varied according to the type of intervention rather than for the clinical specialization. This is illustrated by the strong and statistically significant correlation between the priorities of the neurologist performing elective surgery and the surgeon in charge of ear cochlear implants in children (Table  3). Discussion of the results with the responders confirmed that their needs were the same: first scanner performance (in both cases anatomical details and processing capability were crucial), then patient safety (an issue which is a priority for the whole medical field), usability and finally technical issues (considered important but not as much as the other needs). Table  3 demonstrated that no significant rank correlation was observed between the neurologists performing elective and emergency surgeries. Finally, the rankings between surgeons working in emergency departments were strongly and significantly correlated (Table  3). Discussion of these results with responders confirmed that their needs were the same: first patient safety (due to the unstable condition of the majority of their patients), then performance (execution time was crucial, once again due to patient instability), then usability and finally technical issues. The clinician in charge of the radiology unit ranked the need categories similarly to the emergency surgeons, but with different motivations: first patient safety (as a general medical approach, but also because of legal responsibility), then performance (to address working organization, unit competitiveness and radiologist scientific interest), usability and technical issues.

Regarding local weights within the category of Performance (Table  2), in elective surgeries, spatial resolution was considered the most important need. This reflected the fact that there are similarities between neuro-surgeries and cochlear implantations in terms of the need to investigate small anatomical details. For this type of case, the neurologist considered the processing software almost as important as the spatial resolution, reflecting the fact that the images used for neurology surgery require more complex pre- and post- processing than those for ear implants. Speed was not considered crucial mainly because the patients undergoing this procedure are usually stable. Again regarding the performance, in emergency surgeries, speed run was considered of paramount importance due to the unstable condition of the patients, which placed them at risk of death or serious impairment. The neurologist reported that spatial resolution was as important as speed run, due to the importance of anatomical details in neurosurgery. Processing software was reported as the least important issue as in emergency situations real-time information is crucial and software requires time to process images. The prioritization of the radiologist was more similar to the rankings of emergency surgeries than elective. Once again, by discussing this result with the Trust Medical Director, it emerged that the majority of radiologist activities are requested from the emergency unit and therefore the daily activities of the radiologist influenced his priorities.

Regarding the local weights in the category of Safety (Table  2), in elective surgeries as in radiology, all the issues were considered equally important. This is likely to be due to the fact that patient safety is an important issue in all branches of medicine. However, it takes on even greater importance in emergency situations, and this was reflected by the differences in importance between needs in the safety category for emergency surgery. Patient monitoring was scored as most important, as patients are frequently in unstable conditions during these kinds of surgeries. The neurologist also considered contrast medium control as important as the brain is particularly sensitive to these drugs. Radiation dose was considered less important during emergencies as the critical nature of these procedures justify some risk to the patient from radiation exposure.

The highest variation in local weights of needs was found in the category of Usability (Table  2). This reflects different needs with regard to this factor; the radiologist, the surgeon responsible for cochlear implants and the emergency neurologist scored application support as the most important need. The neurologists considered interoperability important, for both emergency and elective surgeries. This reflected the fact that they often needed to integrate information from images obtained with different technologies (ultrasound, magnetic resonance and CT).

Regarding local weights of needs in the Technical issues category (Table  2), no significant information emerged from the radiologist and the ear surgeon. The elective neurologist considered data storing important. In emergency, technical assistance was considered of paramount importance. Discussing this result with the emergency surgeons revealed that time to first intervention, up time and mean time to repair were considered important to guarantee service continuity. These were not considered crucial for elective surgery, where the number of interventions in the year and the condition of the patients meant that some delays were acceptable.

Regarding global weights, Table  2 shows that for elective surgery the top five important needs are the same: spatial resolution, processing software, radiation dose, patient monitoring, and contrast medium. Similarly, in emergency surgery, the top five needs were the same: patient monitoring, radiation dose, contrast medium control, speed run, spatial resolution. Table  4 shows that there was again a higher rank correlation according to surgery, election-election (86%) or emergency-emergency (90%), more than according to specialization: neurologist-neurologist (ρ<50% and p>0.05). In addition, radiologist prioritization was significantly and strongly correlated to emergency (82% and 77% with p<0.01) more than to elective surgery. In this case, a significant correlation between radiologist and ear surgeon (73%, p<0.1) was observed. This result was unexpected considering that the number of CT scans required for ear-surgery represents less than the 5% of the total activity of radiology. From a methodological point of view, this result was mainly due to the fact that both the radiologist and the clinician responsible for ear-surgery scored all the needs in safety and technical categories as equally important. This could be a weakness of this method. Nonetheless, after discussing this result with the radiologists and with the Medical Director of the Trust, it emerged that this strong correlation was likely to be due to the fact that radiologist and ear surgeons had collaborated in designing surgery for cochlear implants and in this kind of intervention, computer assisted design in pre-surgery planning is crucial to select the cochlear device and to plan the implant. This may illustrate the strength of the AHP method in mapping specific needs of specific trusts.

Regarding the method, it should be noted that AHP is normally used within a group decision-making process and requires that the decision-makers meet to compare and discuss their weights and decisions as a means to develop a consensus on group weights and achieve a group decision. However, this was not the purpose of this study, which aimed instead to explore the differences between the needs of clinicians with different specializations and different clinical settings. We have demonstrated that there was high consensus between those clinicians working in similar settings (emergency versus elective medicine), independent of their clinical specialization. Regarding the usability of the method, all of the responders reported that they encountered no difficulties in completing the questionnaires and that the results accurately reflected their needs. Moreover, all declared that they would not have been able to spontaneously quantify their preferences in such a detailed manner. Furthermore, all five responders declared that the method helped them to elicit their needs. The other domain experts involved in this study found the method clear and useful for facilitating the user needs elicitation process. Limiting the number of elements in each category to three assisted the responders, who were not experienced with this method, particularly in avoiding inconsistency and speeding up the process. The scale used, from 1 to 5 and not to 9 as proposed by Saaty [50], resulted in more significance to responders, as already stated in previous research [26]. This was possible because of the low number of elements in each node. The careful design of the questionnaires facilitated responders’ coherence, which has been identified as an important issue in avoiding inconsistencies by other AHP studies in healthcare [41,42], especially when responders are patients. This is because AHP requires that the words used are familiar to lay responders and therefore care must be taken when naming needs and categories. Although, in this study, the responders were clinicians with extensive experience of the topics and terms under investigation it is still important to reduce the risks of confusion or misunderstanding.

This study supports the results of previous studies [28] that using a limited number of elements in the same node of the hierarchy may reduce inconsistencies. This study confirms the results of previously published papers [54] that less than five elements per node can be considered a satisfactory threshold to achieve a good level of significance. In addition, a reduced number of possible judgments, a 1 to 5 scale instead of a 1 to 9 one, reduced inconsistencies [28].

Therefore, to apply AHP method in a healthcare context, especially when patient and lay users are involved, we recommend: (1) the use of a limited number of possible judgments, for example a 1 to 5 scale, and (2) to put no more than 4 elements in each node. This last recommendation may require a deeper hierarchy, but it has been demonstrated that by adding more levels, the total number of questions is globally reduced [57].

Regarding the limitations of this study, the number of responders was relatively small, which means that it was not possible to investigate whether preferences for CT scanning varied, for example according to factors such as age, length of clinical experience and educational background. In addition, it also means that it is not possible to generalize the results to different scenarios such as different hospitals. Regarding the method, although according to the pyramid of evidence, studies basing on opinions are not considered the most reliable, a gap exist between evidence and every-day decision making healthcare organizations. AHP may contribute to combine empirical evidence and subjective experience in order to improve medical decision-making.

Conclusion

User needs elicitation is a fundamental part of device design and purchasing. The method described in this paper allowed user needs to be elicited according to different working scenarios and medical specializations. Moreover, AHP provided an understandable and traceable framework for the decision process, which is essential in the public sector where decision makers are required to justify their choices to different stakeholders. This paper has demonstrated that, for this case study of a CT scanner, user requirements varied more according to medical scenario (elective surgery versus emergency) than to clinical specialization. This should be considered before when deciding whether to allocate budgets for medical devices according to clinical functions or according to hospital units. These results also have important implications for the manufacturers of CT scanners as they suggest that decisions on device functionality and features should be made according to the medical scenario rather than the clinical specialization. This would then enable manufacturers to produce competitively priced devices, which are appropriate for the particular clinical setting. The study also has wider implications for the medical device industry as it describes a rigorous and effective method for eliciting user requirements during the development of new devices. Finally, when using AHP in healthcare, two issues should be considered: firstly, to use a limited number of items in each node, and secondly, to use a limited scale for responders’ judgment.

Competing interests

The authors declare that they have no competing interests.

Authors’ contribution

LP, AR and LM conceived this study. LP and CV drafted the hierarchy and the questionnaires, analysed the data and presented the results. LP, AR, AS, LM participated to the focus group, reviewed the hierarchy of factors, prepared the ethical application, enrolled the responders, coordinated the elicitation study, submitted the questionnaires, discussed the results with other medical personnel. LP, JLM and SPM discussed the results considering the state of the art of the literature, drafted the paper and reviewed the manuscript. All the authors contributed to the paper. All authors read and approved the final manuscript.

Pre-publication history

The pre-publication history for this paper can be accessed here:

http://www.biomedcentral.com/1472-6947/13/2/prepub

Contributor Information

Leandro Pecchia, Email: leandro.pecchia@nottingham.ac.uk.

Jennifer L Martin, Email: jennifer.martin@nottingham.ac.uk.

Angela Ragozzino, Email: angelarag56@virgilio.it.

Carmela Vanzanella, Email: carmela.vanzanella@cnr.it.

Arturo Scognamiglio, Email: ascognamiglio@hotmail.it.

Luciano Mirarchi, Email: luciano.mirarchi@siemens.com.

Stephen P Morgan, Email: steve.morgan@nottingham.ac.uk.

Acknowledgements

LP, JLM and SPM acknowledge support of this work through the MATCH Programme (EPSRC Grant EP/F063822/1) although the views expressed are entirely their own.

References

  1. Sawyer D. Do it by design. An introduction to human factors in medical devices. http://www.fda.gov/medicaldevices/deviceregulationandguidance/guidancedocuments/ucm094957.htm
  2. Martin JL, Murphy E, Crowe JA, Norris BJ. Capturing user requirements in medical device development: the role of ergonomics. Physiol Meas. 2006;27(8):R49–R62. doi: 10.1088/0967-3334/27/8/R01. [DOI] [PubMed] [Google Scholar]
  3. Shah SGS, Robinson I. Benefits of and barriers to involving users in medical device technology development and evaluation. Int J Technol Assess Health Care. 2007;23(1):131–137. doi: 10.1017/S0266462307051677. [DOI] [PubMed] [Google Scholar]
  4. Money AG, Barnett J, Kuljis J, Craven MP, Martin JL, Young T. The role of the user within the medical device design and development process: medical device manufacturers’ perspectives. BMC Med Inform Decis Mak. 2011;11:15. doi: 10.1186/1472-6947-11-15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Martin JL, Barnett J. Integrating the results of user research into medical device development: insights from a case study. BMC Med Inform Decis Mak. 2012;12:74. doi: 10.1186/1472-6947-12-74. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Bracale U, Rovani M, Picardo A, Merola G, Pignata G, Sodo M, Di Salvo E, Ratto EL, Noceti A, Melillo P, Beneficial effects of fibrin glue (Quixil) versus Lichtenstein conventional technique in inguinal hernia repair: a randomized clinical trial. Hernia. 2012. Epub ahead of print. [DOI] [PubMed]
  7. Bracale U, Rovani M, Bracale M, Pignata G, Corcione F, Pecchia L. Totally laparoscopic gastrectomy for gastric cancer: Meta-analysis of short-term outcomes. Minim Invasive Ther Allied Technol. 2011;21(3):150–160. doi: 10.3109/13645706.2011.588712. [DOI] [PubMed] [Google Scholar]
  8. Bracale U, Rovani M, Melillo P, Merola G, Pecchia L. Which is the best laparoscopic approach for inguinal hernia repair: TEP or TAPP? A network meta-analysis. Surg Endosc. 2012. Epub ahead of print. [DOI] [PubMed]
  9. Leys M. Health care policy: qualitative evidence and health technology assessment. Health Policy. 2003;65(3):217–226. doi: 10.1016/S0168-8510(02)00209-9. [DOI] [PubMed] [Google Scholar]
  10. Martin JL, Norris BJ, Murphy E, Crowe JA. Medical device development: The challenge for ergonomics. Appl Ergon. 2008;39(3):271–283. doi: 10.1016/j.apergo.2007.10.002. [DOI] [PubMed] [Google Scholar]
  11. Pope C, Ziebland S, Mays N. Qualitative research in health care - Analysing qualitative data (Reprinted from Qualitative Research in Health Care) Br Med J. 2000;320(7227):114–116. doi: 10.1136/bmj.320.7227.114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Upshur REG, VanDenKerkhof EG, Goel V. Meaning and measurement: an inclusive model of evidence in health care. J Eval Clin Pract. 2001;7(2):91–96. doi: 10.1046/j.1365-2753.2001.00279.x. [DOI] [PubMed] [Google Scholar]
  13. Kaplan B, Shaw NT. Future directions in evaluation research: People, organizational, and social issues. Methods Inf Med. 2004;43(3):215–231. [PubMed] [Google Scholar]
  14. Sackett DL, Haynes RB. Evidence base of clinical diagnosis - The architecture of diagnostic research. Br Med J. 2002;324(7336):539–541. doi: 10.1136/bmj.324.7336.539. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Malterud K. Qualitative research: standards, challenges, and guidelines. Lancet. 2001;358(9280):483–488. doi: 10.1016/S0140-6736(01)05627-6. [DOI] [PubMed] [Google Scholar]
  16. Chapple A, Rogers A. Explicit guidelines for qualitative research: a step in the right direction, a defence of the ‘soft’ option, or a form of sociological imperialism? Fam Pract. 1998;15(6):556–561. doi: 10.1093/fampra/15.6.556. [DOI] [PubMed] [Google Scholar]
  17. Hostgaard AM, Bertelsen P, Nohr C. Methods to identify, study and understand End-user participation in HIT development. BMC Med Inform Decis Mak. 2011;11:57. doi: 10.1186/1472-6947-11-57. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Cios KJ, Moore GW. Uniqueness of medical data mining. Artif Intell Med. 2002;26(1–2):1–24. doi: 10.1016/s0933-3657(02)00049-0. [DOI] [PubMed] [Google Scholar]
  19. Melillo P, Fusco R, Sansone M, Bracale M, Pecchia L. Discrimination power of long-term heart rate variability measures for chronic heart failure detection. Med Bio Eng Comput. 2011;49(1):67–74. doi: 10.1007/s11517-010-0728-5. [DOI] [PubMed] [Google Scholar]
  20. Melillo P, Izzo R, Luca N, Pecchia L. Heart rate variability and target organ damage in hypertensive patients. BMC Cardiovasc Disord. 2012;12(1):105. doi: 10.1186/1471-2261-12-105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Pecchia L, Mirarchi L, Doniacovo R, Marsico V, Bracale M. Health Technology Assessment for a Service Contract: a new method for decisional tools. World Congress on Medical Physics and Biomedical Engineering. 2009;25(12):105–108. [Google Scholar]
  22. Benario HW. Caesar’s Gallic war: a commentary. Norman: University of Oklahoma Press; 2012. [Google Scholar]
  23. Raible F, Brand M. Divide et Impera–the midbrain-hindbrain boundary and its organizer. Trends Neurosci. 2004;27(12):727–734. doi: 10.1016/j.tins.2004.10.003. [DOI] [PubMed] [Google Scholar]
  24. Scorrano L. Divide et impera: Ca2+ signals, mitochondrial fission and sensitization to apoptosis. Cell Death Differ. 2003;10(12):1287–1289. doi: 10.1038/sj.cdd.4401310. [DOI] [PubMed] [Google Scholar]
  25. Reinhardt U. Divide et impera: protecting the growth of health care incomes (COSTS) Health Econ. 2012;21(1):41–54. doi: 10.1002/hec.1813. [DOI] [PubMed] [Google Scholar]
  26. Pecchia LB, P A, Pendleton N, Jackson S, Clarke C, Briggs P, Mcinnes L, Angelova M, Bracale M. The use of analytic hierarchy process for the prioritization of factors affecting wellbeing in elderly. Sorrento, Naples, Italy; 2011. pp. 1–4. (proceedings of the 11th. International Symposium on Analytic Hierarchy Process (ISAHP)). [Google Scholar]
  27. Saaty TL. An essay on how judgment and measurement are different in science and in decision making. International Journal of the Analytic Hierarchy Process. 2009;1(1):61–62. [Google Scholar]
  28. Pecchia L, Bath PA, Pendleton N, Bracale M. Analytic Hierarchy Process (AHP) for examining healthcare professionals’ assessments of risk factorsThe relative importance of risk factors for falls in community-dwelling older people. Methods Inf Med. 2011;50(5):435–444. doi: 10.3414/ME10-01-0028. [DOI] [PubMed] [Google Scholar]
  29. Hummel JM, IJzerman MJ. A Systematic Review of the Analytic Hierarchy Process in Health Care Decision Making. Value Health. 2009;12(7):A227–A227. [Google Scholar]
  30. Uzoka FM, Obot O, Barker K, Osuji J. An experimental comparison of fuzzy logic and analytic hierarchy process for medical decision support systems. Comput Methods Programs Biomed. 2011;103(1):10–27. doi: 10.1016/j.cmpb.2010.06.003. [DOI] [PubMed] [Google Scholar]
  31. Danner M, Hummel JM, Volz F, van Manen JG, Wiegard B, Dintsios C-M, Bastian H, Gerber A, Ijzerman MJ. Integrating patients’ views into health technology assessment: Analytic hierarchy process (AHP) as a method to elicit patient preferences. Int J Technol Assess Health Care. 2011;27(4):369–375. doi: 10.1017/S0266462311000523. [DOI] [PubMed] [Google Scholar]
  32. Bridges JF. Future challenges for the economic evaluation of healthcare: patient preferences, risk attitudes and beyond. PharmacoEconomics. 2005;23(4):317–321. doi: 10.2165/00019053-200523040-00002. [DOI] [PubMed] [Google Scholar]
  33. Dolan JG. Are patients capable of using the analytic hierarchy process and willing to use it to help make clinical decisions? Med Decis Making. 1995;15(1):76–80. doi: 10.1177/0272989X9501500111. [DOI] [PubMed] [Google Scholar]
  34. Dolan JG. Multi-criteria clinical decision support: A primer on the use of multiple criteria decision making methods to promote evidence-based, patient-centered healthcare. Patient. 2010;3(4):229–248. doi: 10.2165/11539470-000000000-00000. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Bridges JF, Carswell CI. Andrew lloyd: a driving force in patient-centered outcomes research. Patient. 2008;1(4):259–263. doi: 10.2165/1312067-200801040-00007. [DOI] [PubMed] [Google Scholar]
  36. Bridges JF, Mohamed AF, Finnern HW, Woehl A, Hauber AB. Patients’ preferences for treatment outcomes for advanced non-small cell lung cancer: A conjoint analysis. Lung Cancer. 2012;77(1):224–231. doi: 10.1016/j.lungcan.2012.01.016. [DOI] [PubMed] [Google Scholar]
  37. de Bekker-Grob EW, Ryan M, Gerard K. Discrete choice experiments in health economics: a review of the literature. Health Econ. 2012;21(2):145–172. doi: 10.1002/hec.1697. [DOI] [PubMed] [Google Scholar]
  38. Gallego G, Bridges JF, Flynn T, Blauvelt BM. Predicting the Future Impact of Emerging Technologies on Hepatocellular Carcinoma (Hcc): Measuring Stakeholders Preferences with Best-Worst Scaling. Value Health. 2011;14(3):A176–A176. [Google Scholar]
  39. Scholl A, Manthey L, Helm R, Steiner M. Solving multiattribute design problems with analytic hierarchy process and conjoint analysis: An empirical comparison. Eur J Oper Res. 2005;164(3):760–777. doi: 10.1016/j.ejor.2004.01.026. [DOI] [Google Scholar]
  40. Mulye R. An empirical comparison of three variants of the AHP and two variants of conjoint analysis. J Behav Decis Mak. 1998;11(4):263–280. doi: 10.1002/(SICI)1099-0771(1998120)11:4&#x0003c;263::AID-BDM301&#x0003e;3.0.CO;2-T. [DOI] [Google Scholar]
  41. Ijzerman MJ, van Til JA, Snoek GJ. Comparison of two multi-criteria decision techniques for eliciting treatment preferences in people with neurological disorders. Patient. 2008;1(4):265–272. doi: 10.2165/1312067-200801040-00008. [DOI] [PubMed] [Google Scholar]
  42. Ijzerman MJ, van Til JA, Bridges JF. A comparison of analytic hierarchy process and conjoint analysis methods in assessing treatment alternatives for stroke rehabilitation. Patient. 2012;5(1):45–56. doi: 10.2165/11587140-000000000-00000. [DOI] [PubMed] [Google Scholar]
  43. Liberatore MJ, Nydick RL. The analytic hierarchy process in medical and health care decision making: A literature review. Eur J Oper Res. 2008;189(1):194–207. doi: 10.1016/j.ejor.2007.05.001. [DOI] [Google Scholar]
  44. Tarimcilar MM, Khaksari SZ. Capital-budgeting in Hospital Management using the analytic hierarchy process. Socioecon Plann Sci. 1991;25(1):27–34. doi: 10.1016/0038-0121(91)90026-N. [DOI] [PubMed] [Google Scholar]
  45. van Til JA, Renzenbrink GJ, Dolan JG, Ijzerman MJ. The use of the analytic hierarchy process to aid decision making in acquired equinovarus deformity. Arch Phys Med Rehabil. 2008;89(3):457–462. doi: 10.1016/j.apmr.2007.09.030. [DOI] [PubMed] [Google Scholar]
  46. Hu H. Multi-slice helical CT: scan and reconstruction. Med Phys. 1999;26(1):5–18. doi: 10.1118/1.598470. [DOI] [PubMed] [Google Scholar]
  47. Kroft J, Klostermann NR, Moody JRK, Taerk E, Wolfman W. A novel regimen of combination transdermal estrogen and intermittent vaginally administered progesterone for relief of menopausal symptoms. Gynecol Endocrinol. 2010;26(12):902–908. doi: 10.3109/09513590.2010.487602. [DOI] [PubMed] [Google Scholar]
  48. Pecchia L, Bracale U, Bracale M. Health Technology Assessment of Home Monitoring for the Continuity of Care of patient suffering from congestive heart failure. World Congress on Medical Physics and Biomedical Engineering. 2009;25(12):184–187. [Google Scholar]
  49. Saaty TL, Vargas LG. Models, methods, concepts & applications of the analytic hierarchy process. Boston: Kluwer Academic Publishers; 2001. [Google Scholar]
  50. Saaty TL. A scaling method for priorities in hierarchical structures. J Math Psychol. 1977;15:8. [Google Scholar]
  51. Ji P, Jiang R. Scale transitivity in the AHP. J Oper Res Soc. 2003;54(8):896–905. doi: 10.1057/palgrave.jors.2601557. [DOI] [Google Scholar]
  52. Finan JS, Hurley WJ. Transitive calibration of the AHP verbal scale. Eur J Oper Res. 1999;112(2):367–372. doi: 10.1016/S0377-2217(97)00411-6. [DOI] [Google Scholar]
  53. Salo AA, Hamalainen RP. The measurement of preferences in the analytic hierarchy process. J Multi-Criteria Decis Anal. 1997;6:11. doi: 10.1002/(SICI)1099-1360(199701)6:1&#x0003c;11::AID-MCDA113&#x0003e;3.0.CO;2-K. [DOI] [Google Scholar]
  54. Pecchia L, Bath PA, Pendleton N, Bracale M. Web-based system for assessing risk factors for falls in community-dwelling elderly people using the analytic hierarchy process. International Journal of the Analytic Hierarchy Process. 2010;2(2):135–157. [Google Scholar]
  55. Pecchia L, Bath P, Pendleton N, Bracale M. In: Proceedings of the 10th International Symposium on AHP (ISAHP2009): July 29–August 1. Tammy T, editor. Pennsylvania, USA: University of Pittsburgh, Pittsburgh; 2009. AHP and risk management: a case study for assessing risk factors for falls in community-dwelling older patients; pp. 1–15. ISSN 1556-8296. [Google Scholar]
  56. Pecchia L, Bracale U, Melillo P, Sansone M, Bracale M. In: Proceedings of the 10th International Symposium on AHP (ISAHP2009): July 29–August 1. Tammy T, editor. Pennsylvania, USA: University of Pittsburgh, Pittsburgh; 2009. AHP for Health Technology Assessment. A case study: prioritizing care approaches for patients suffering from chronic heart failure; pp. 1–9. ISSN 1556-8296. [Google Scholar]
  57. Saaty T. How to Structure and Make Choices in Complex Problems. Hum Syst Manag. 1982;3(4):255–261. [Google Scholar]
  58. Uzoka FME. A fuzzy-enhanced multicriteria decision analysis model for evaluating university Academics’ research output. Information Knowledge Systems Management. 2008;7 [Google Scholar]
  59. Carmone FJ, Kara A, Zanakis SH. A Monte Carlo investigation of incomplete pairwise comparison matrices in AHP. Eur J Oper Res. 1997;102(3):538–553. doi: 10.1016/S0377-2217(96)00250-0. [DOI] [Google Scholar]

Articles from BMC Medical Informatics and Decision Making are provided here courtesy of BMC

RESOURCES