Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2020 Feb 27;20(5):1283. doi: 10.3390/s20051283

Automatic, Qualitative Scoring of the Interlocking Pentagon Drawing Test (PDT) Based on U-Net and Mobile Sensor Data

Ingyu Park 1, Yun Joong Kim 2, Yeo Jin Kim 3,*, Unjoo Lee 1,*
PMCID: PMC7085787  PMID: 32120879

Abstract

We implemented a mobile phone application of the pentagon drawing test (PDT), called mPDT, with a novel, automatic, and qualitative scoring method for the application based on U-Net (a convolutional network for biomedical image segmentation) coupled with mobile sensor data obtained with the mPDT. For the scoring protocol, the U-Net was trained with 199 PDT hand-drawn images of 512 × 512 resolution obtained via the mPDT in order to generate a trained model, Deep5, for segmenting a drawn right or left pentagon. The U-Net was also trained with 199 images of 512 × 512 resolution to attain the trained model, DeepLock, for segmenting an interlocking figure. Here, the epochs were iterated until the accuracy was greater than 98% and saturated. The mobile senor data primarily consisted of x and y coordinates, timestamps, and touch-events of all the samples with a 20 ms sampling period. The velocities were then calculated using the primary sensor data. With Deep5, DeepLock, and the sensor data, four parameters were extracted. These included the number of angles (0–4 points), distance/intersection between the two drawn figures (0–4 points), closure/opening of the drawn figure contours (0–2 points), and tremors detected (0–1 points). The parameters gave a scaling of 11 points in total. The performance evaluation for the mPDT included 230 images from subjects and their associated sensor data. The results of the performance test indicated, respectively, a sensitivity, specificity, accuracy, and precision of 97.53%, 92.62%, 94.35%, and 87.78% for the number of angles parameter; 93.10%, 97.90%, 96.09%, and 96.43% for the distance/intersection parameter; 94.03%, 90.63%, 92.61%, and 93.33% for the closure/opening parameter; and 100.00%, 100.00%, 100.00%, and 100.00% for the detected tremor parameter. These results suggest that the mPDT is very robust in differentiating dementia disease subtypes and is able to contribute to clinical practice and field studies.

Keywords: pentagon drawing test, automatic scoring, mobile sensor, deep learning, U-Net, Parkinson’s disease

1. Introduction

The pentagon drawing test (PDT) is a sub-test of the Mini-Mental State Examination (MMSE), used extensively in clinical and research settings as a measure of cognitive impairment [1]. The MMSE is a general screening tool for cognitive impairment. However it shows low sensitivity for detecting cognitive impairment in Parkinson’s disease [2].

Parkinson’s disease is the second most common neurodegenerative disease and is characterized by motor symptoms [3]. In addition, 83% of long-term survivors of Parkinson’s disease showed dementia, with impairment of their visuospatial functions and executive functions [4]. As up to half of Parkinson’s patients show visuospatial dysfunction, the PDT has also been used for the distinction of dementia in Parkinson’s disease and Alzheimer’s disease cases [5,6,7,8,9,10]. However, the conventional PDT scoring method shows a low sensitivity in detecting visuospatial dysfunction in Parkinson’s disease. It is necessary to improve the scoring method of the PDT to measure visuospatial dysfunction more accurately.

Visuospatial functioning includes various neurocognitive abilities, such as coordinating fine motor skills, identifying visual and spatial relationship among objects, and planning and executing tasks [9]. These functions allow us to estimate distance and perceive depth, use mental imagery, copy drawings, and construct objects or shapes [11]. Before clinical symptoms of Parkinson’s or Alzheimer’s disease become manifest, various visuospatial dysfunctions are known to be detectable for several years before, thus indicating that visual spatial assessment can assist in making an early and accurate diagnosis of these cases [12].

There are several tests including the clock drawing test (CDT) and the Rey–Osterrieth complex figure (ROCF) test, in addition to the PDT, for measuring visuospatial function. The CDT is applied for early screening of cognitive impairment, especially in dementia. It can be used to assess tasks, such as following directions, comprehending language, visualizing the proper orientation of an object, and executing normal movements. These may be disturbed in dementia [13,14]. Due to the dominant roles of the CDT in dementia, different types of scoring systems are relatively well established, ranging from semi-quantitative to qualitative scoring systems [15,16]. There are also several suggested methods for performing the CDT in a digital environment and/or scoring it based on machine learning, convolutional neural networks, and ontology representations [17,18,19,20].

The ROCF test employs a complex geometrical figure for the stimulus, comprised of a large rectangle with horizontal and vertical bisectors, two diagonals, and additional geometric details [21]. The ROCF is widely used for testing visual perception and long-term visual memory with its usefulness in providing information on the location and extent of any damage, and through the order and accuracy in which the figure is copied and drawn from recall [22]. Various scoring systems for the ROCF test, including the Osterrieth system and the Boston Qualitative scoring system, have been employed to derive a more quantitative value for the accuracy of a subject’s drawing. However, these quantify the accuracy of the drawings by hand and in a subjective manner due to the complexity of the figures [23]. To address this, an automated scoring algorithm for the ROCF test was recently suggested and is based on cascaded deep neural networks, trained on scores from human expert raters [24].

As for the PDT, it is a relatively fast, simple, and sensitive test compared to the CDT and the ROCF tests. For the test, a subject is asked to copy or draw two interlocking pentagons with the interlocking shape being a rhombus. The interpretation of PDT is usually binary, but there are several ways of interpreting PDT such as a feasible scoring for a total of 6 or 10 points [25,26], a standardized scoring for use in psychiatric populations [27], and a qualitative measure with a total of 13 points [5]. The established scores for a total of 6 points are as follows: 6 points for correct copying; 5 points for two overlapping pictures, one being a pentagon; 4 points for only two overlapping pictures; 3 points for having two figures, not overlapping; 2 points for a closed figure; and 1 point when the drawing does not have the shape of a closed figure. The differences in the length of the sides were additionally considered in the scoring, giving an overall total of 10 points. The standardized scoring increased the sensitivity and reliability for use in nonorganic psychiatric patients by also qualitatively considering the distortion degree of the pentagon shapes. The qualitative scoring involved the assessment of drawing parameters, including the number of angles, distance/intersection, closure/opening, rotation, and closing-in, in the hand-drawn images for the PDT.

Many studies have shown evidence for the PDT being prognostic for the assessment of visuospatial functions. The PDT has been used in the differentiation of dementia with Lewy Bodies, Alzheimer’s disease, Parkinson’s disease (with or without dementia), paraphrenia, schizophrenia, and obsessive compulsive disorder, with the qualitative scoring systems applied rather than using the binary scoring system [6,9,26,27,28,29,30]. An associative study of the PDT and the CDT has also shown the PDT to be applicable as a prognostic marker in dementia with Lewy Bodies [31]. From these studies, the qualitative scoring of the PDT is practical and effective in distinguishing several clinical features of various cognitive deficits. However, as the qualitative scoring of the PDT is done manually, it can be prone to human error and is not very practical in analyzing big data, such as for personal lifelogs. With the PDT result being qualitative and not numeric, the results are also difficult to evaluate objectively. As such, the necessity of a standardized, qualitative, and automatic scoring system for the PDT has increased. The availability of digital and mobile sensors, coupled with deep learning algorithms, has made it possible to think of ways to using this technology to obtain information for the PDT and to add order and accuracy for interpreting PDT figures as they are being copied and drawn from recall. This may provide additional information on the location and extent of any damage.

In this study, we developed a mobile-phone application version of the PDT, named the mPDT, and a novel automatic, qualitative scoring system, based on U-Net, a convolutional network for biomedical image segmentation (see Appendix A for the summary and comparison of this study to those of other reports). The mobile sensor data from a smart phone allowed for development of this mPDT application and the scoring system. The U-Net was trained with 199 PDT hand-drawn images of 512 × 512 resolution obtained via mPDT to generate a trained model, Deep5, in segmenting the drawn right and left pentagon images. The U-Net was also trained with the same 199 PDT 512 × 512 resolution images for DeepLock, also a trained model, for segmenting an interlocking image of two pentagons.

The epochs were then iterated for both Deep5 and DeepLock until the accuracies were greater than 98% and saturated as well. The mobile sensor data consisted of primary and secondary data, where the primary data were the x and y coordinates, timestamps, and touch events for all the samples with a 20 ms sampling period extracted from the mobile touch sensor. The secondary data were the velocities calculated using the primary data. Four parameters, including the number of angles (0–4 points), distance/intersection between the two figures (0–4 points), closure/opening of the contour (0–2 points), and detected tremors (0–1 points), were estimated using Deep5 and DeepLock and the sensor data, resulting in scaling with a total of 11 points.

The performance test was performed with images and sensor data from 230 subjects, obtained via the mPDT. All the images were scored by two clinical experts in PDT scaling for an objective performance analysis. The results of the performance test indicated the sensitivity, specificity, accuracy, and precision for the number of angles at 97.53%, 92.62%, 94.35%, and 87.78%; for the distance/intersection at 93.10%, 97.90%, 96.09%, and 96.43%; for the closure/opening at 94.03%, 90.63%, 92.61%, and 93.33%; and for tremors at 100.00%, 100.00%, 100.00%, and 100.00%, respectively. These results suggest that our mPDT application is valuable in differentiating dementia disease subtypes and also useful for clinical practice and field studies.

2. Materials and Methods

2.1. Subjects

The Institutional Review Board of the Hallym University Sacred Heart Hospital approved this study. A total of 328 right-handed young volunteers (175 males and 153 females, aged (mean ± std.) 23.98 ± 2.83 years) were recruited and participated in the pentagon drawing test using the mPDT, our mobile application developed in this study. In addition, 101 pentagon drawing images by Parkinson’s disease (PD) patients (47 males and 54 females, aged (mean ± std.) 75.86 ± 8.13 years) that visited the university hospital were used in this study. The pentagon drawing image data of 199 volunteers (107 males and 92 females, aged (mean ± std.) 22.11 ± 1.44 years) from the aforementioned 328 volunteers were used in creating the pre-training models of Deep5 and DeepLock, using the U-Net algorithm. The remaining 129 volunteers (68 males and 61 females, aged (mean ± std.) 26.87 ± 1.84 years) provided the pentagon drawing image data, along with those from the 101 PD patients, that were used in testing the scoring method. All the images were scored by two clinical experts in PDT scaling. Table 1 summarizes certain informative statistics of age, gender, and binary PDT score of the 328 volunteers and 101 PD patients.

Table 1.

Statistics of age, gender, handedness, and clinical status of the participants. Pentagon drawing test (PDT).

Training Set Test Set
Volunteers (n = 199) Volunteers (n = 129) PD Patients (n = 101)
Age (mean ± standard deviation) 22.11 ± 1.44 26.87 ± 1.84 75.86 ± 8.13
Gender (male/female) 107/92 68/61 47/54
Binary PDT score (Pass/Nonpass ) 199/0 48/81 32/69

2.2. Implementation of the Mobile Pentagon Drawing Test, mPDT

The mobile application, mPDT, for the interlocking pentagon test was developed using the Android Studio development environment. While the source code of mPDT was implemented to be able to be built in any mobile device, including smartphones, tablets, or notebooks, we built and tested the mPDT in a Samsung Galaxy Note 4 smartphone with a resolution of 640 dots per inch and a spatial accuracy of 0.004 cm. The mPDT allows for a user to copy two interlocking pentagons (with the interlocking shape being a rhombus) on the screen, and scores the drawing image qualitatively based on the sensor data of the drawing image and the pre-trained models, Deep5 and DeepLock, developed in this study. Using U-Net, a convolutional network architecture for fast and precise segmentation of images, the pre-trained models Deep5 and DeepLock were generated for segmenting the pentagon shapes and the interlocking shape, respectively.

The sensor data collected by the touch sensors embedded in a smartphone consisted of timestamps in seconds, the x and y coordinates in pixels, and touch-events of the samples of the drawing image with a 50 Hz sampling frequency. Figure 1a shows the flow diagram of the mPDT operation in, and Figure 1b–d shows the screen shots of the registration window, the PDT window, and the result window for mPDT, respectively. At the launch of mPDT, an informed consent prompt appears, following this, a registration window is displayed, where it is possible to enter the subject’s information, such as name, age, gender, and handedness, plus the optional parameters.

Figure 1.

Figure 1

(a) Flow diagram of the mobile phone pentagon drawing test (PDT) (mPDT) operation; screen shots of the (b) registration window; (c) PDT window; and (d) results window of the mPDT.

Following the user pressing the start button in the registration window, the PDT window then appears, in which the user is asked to copy two interlocking pentagons provided on a paper by an examiner or draw them while recalling from an image provided on a previous window. Values of measured parameters for the time in seconds and the x and y coordinates in pixels are obtained from the drawing image with a 50 Hz sampling frequency. The values are saved as sensor data when the subject copies or draws pentagons on the touch screen of the PDT window. The results window then displays the sensor data along with the drawn image, and/or a plot of speeds in mm/sec of inter-samples over time. The results could then be sent to the email address entered at the registration window.

2.3. Pre-Trained Models, Deep5 and DeepLock based on the U-Net

Novel pre-trained models of Deep5 and DeepLock were developed for segmentation of the drawn pentagon portions and the interlocking domains of the images, respectively. Deep5 and DeepLock were created based on the U-Net convolutional network architecture in keras [32]. The network architecture implemented in this study is illustrated in Figure 2. It consists of a contracting path, an expansive path, and a final layer. The contracting path consists of repeated applications of two 3 × 3 convolutions and a 2 × 2 max pooling operation with stride 2 for down-sampling. At each repetition, the number of feature channels is doubled. The expansive path consists of two 3 × 3 convolutions and a 2 × 2 convolution (“up-convolution”) for up-sampling to recover the size of the segmentation map. At the final layer, a 1 × 1 convolution was used to map each 16-component feature vector to the desired number of classes. In total, the network has 23 convolutional layers. The training data for both Deep5 and DeepLock contain 960 images of 128 × 128 resolution, which were augmented using a module called ImageDataGenerator in keras.preprocessing.image and resized from the original 199 images of 1600 × 1320 resolution. Deep5 and DeepLock were generated by training the network architecture for five and seven epochs with accuracies of approximately 0.977 and 0.979, respectively. The loss function used for the training was essentially binary cross entropy.

Figure 2.

Figure 2

The U-Net network architecture used in this study.

2.4. Scoring Method of mPDT

The novel, automatic, and qualitative scoring method for the mPDT was developed based on the sensor data and the pre-trained models, Deep5 and DeepLock. Four parameters were included in the scoring method: the number of angles (0–4 points), distance/intersection between the two figures (0–4 points), closure/opening of the image contour (0–2 points), and detected tremors (0–1 points). All the assigned scores for the parameters are integers. A total score corresponding to the sum of individual scores of each parameter ranged from 0 to 11. The parameters, number of angles (0–4 points), distance/intersection between the two figures (0–4 points), and closure/opening of the image contour (0–2 points) were adopted from a previous study by Paolo Caffarra et al. [5]. When a subject executes more than one copy of the pentagons, the last copy is then scored. A detailed list of the parameters used is presented in Table 2 and the overall flowchart and the schematic diagram of the scoring method are shown in Figure 3a,b, respectively.

Table 2.

A detailed list of the parameters for the scoring method.

Parameters Performance Scores Assigned Integer Scores
Number of Angles 10 4
9 or 11 3
8 or 12 2
7–5 1
< 5 or > 13 0
Distance/Intersection Correct intersection 4
Wrong intersection 3
Contact without intersection 2
No contact, distance < 1 cm 1
No contact, distance > 1 cm 0
Closure/Opening Closure both figures 2
Closure only one figure 1
Opening both figures 0
Tremor No tremor 1
Tremor 0
Total 0–11

Figure 3.

Figure 3

(a) Overall flowchart of the scoring method; (b) Schematic diagram of the scoring method.

The scoring method consists of a series of processes that include manipulation of the sensor data and segmentation of the drawn pentagon and the interlocking shapes using Deep5 and DeepLock, respectively. There is then the extraction of variables (the number of angles, distance/intersection, closure/opening, and presence of tremors) and the assignment of scores according to the performance scores for each parameter. The sensor data primary obtained from the drawn image samples during the mPDT interaction include timestamps, x- and y-coordinates, and touch events. The time index of moving from a figure drawing to another is detected from the primary sensor data in the process of the manipulation of the sensor data. Velocity values are also obtained from the primary sensor data in the process of the manipulation of sensor data.

The drawn image is then analyzed by Deep5 and DeepLock to respectively segment the pentagon and the interlocking shapes. It is here that the percentages of the segmented pentagon shapes and the interlocking shape overlapping the drawn image are calculated. Next, the values for each parameter are calculated. For the total number of angles of each figure, the percentages and the number of peaks in velocities are determined. For the distance/intersection, the interlocking distance between two figures and the absolute ratio of the differences in x and y coordinates of the two points used in the calculation of the interlocking distance are extracted. After clustering the primary sensor data of zero velocities, the mean distance between the cluster center and the points belonging to each cluster are calculated for the closure/opening parameter. For presence of tremors, the frequency of consecutive ‘up and down’ touch events from the sensor data is estimated. Finally, there is the assignment of integer scores according to each parameter score, as in Table 2 (also see Appendix B for the examples of drawn images and the corresponding performance scores and assigned integer scores). The details are described in the following subsections.

2.4.1. Sensor Data Manipulation and Shape Segmentation Using Deep5 and Deeplock

Once the mPDT has been completed, three pieces of data, including the drawing image and the primary and secondary sensor data, are generated for output. The size of the drawing image ID is 128 × 128 pixels, which is resized from the original 1600 × 1320 drawing image of the PDT window. The primary sensor data consist of times t[n] in seconds, x- and y- coordinates, x[n] and y[n] in pixels, and touch-events e[n] of the sample points of the 128 × 128 drawing image, where the sampling rate was set at 50 Hz and n is the index of a sample point. The secondary sensor data are velocities v[n] in pixels/sec which are calculated from the primary sensor data. Each of the touch-events e[n] has one of the values, such as -1, 0, 1, where the assigned value of -1 is for the event ‘down’, as in touching on the screen; 1 for the event ‘up’, as in touching off the screen; and 0 for the event ‘move’, as in moving and touching on the screen. For the drawing image ID, it is supposed to be of two interlocking pentagons.

The sensor data for each of the two interlocking shapes is separately obtained using the times t[nf1] of the last sample point of the first drawn figure and t[nf2] of the first sample point of the second drawn figure. In other words, the sample points for t[n], nnf1 belong to the first figure ID,nnf1 and those for t[n], nnf2 to the second figure ID,nnf2. The time t[nf1] can be estimated by the touch-events shifting from the event ‘move’ into the event ‘up’ and then successively staying at the event ‘up’ for the longest time if such a period occurs more than once. Therefore, the index nf1 can be determined by finding the longest chain of events consisting of a 0 (the event ‘move’) and consecutive 1s (the ‘up’ events) in the sequence of touch-events e[n]. The time t[nf2] can be obtained by adding the longest period of time to the time t[nf1]. Sensor data for the interlocking image IDL of the 128 × 128 drawing image ID is obtained from the sample points between two data points, (x[n1],y[n1]) and (x[n2],y[n2]), where n1 is the index at which the value of x[n1] is the maximum of those belonging to x[n] (n<nf1) and n2 is the index at which the value of x[n2] is the minimum of those belonging to x[n] (nnf2).

The two pentagon shapes Ifi(i=1,2) are then separately segmented using indices nfi(i=1,2) and the Deep5 pre-trained model from the 128 × 128 drawing image ID. An interlocking shape IL is segmented using the DeepLock pre-trained model from the 128 × 128 drawing image ID as well. Next, percentage pfi(i=1,2) of the segmented image Ifi(i=1,2) matching to the corresponding figure of the 128 × 128 drawing image ID is estimated as below:

pf1=n(ID,nnf1If1)n(ID,nnf1),and
pf2=n(ID,nnf2If2)n(ID,nnf2),

where n(ID,nnf1If1) is the number of pixel coordinates that If1 and the first figure of the 128 × 128 drawing image ID have in common; n(ID,nnf1) is the total number of pixel coordinates in the first figure of the 128 × 128 drawing image ID. Similarly, n(ID,nnf2If2) is the number of pixel coordinates that If2 and the second figure of the 128 × 128 drawing image ID have in common; n(ID,nnf2) is the total number of pixel coordinates in the second figure of the 128 × 128 drawing image ID. The percentage pL of the image IL matching to the interlocking image IDL of the 128 × 128 drawing image ID is estimated as follows:

pL=n(IDLIL)n(IDL),

where n(IDLIL) is the number of pixel coordinates that IL and IDL have in common; n(IDL) is the total number of pixel coordinates in IDL.

2.4.2. Number of Angles

Number of angles is the sum of the number NAi (i=1,2) of angles of each figure of the 128 × 128 drawing image ID, which is estimated using the percentage, pfi (i=1,2), velocity v[n], and the index nfi(i=1,2) as follows:

NAi={5pfi>pth1Npielse,

where Npi (i=1,2) is the number of peaks in velocity v[n] for each figure. If the percentage pfi(i=1,2) of the segmented image Ifi(i=1,2) matching to the corresponding figure of the 128 × 128 drawing image ID is larger than a given threshold pth1, then the shape of the figure is identified as a pentagon and the number NAi (i=1,2) of angles is estimated to be 5. If not, then the number NA1 and NA2 are estimated by the numbers Np1 and Np2 of peaks in velocity v[n] (nnf1) and v[n] (nnf2), respectively, since the velocity of the drawing stroke increases and then decreases to zero at the point of change in direction.

2.4.3. Distance/Intersection between Two Figures

The distance/intersection between two figures is evaluated by the percentage pL of the image IL matching to the interlocking image IDL of the 128 × 128 drawing image ID, the distance dL between two figures in the interlocking image, and the ratio Δx/Δy of differences of two data points (x[n1],y[n1]) and (x[n2],y[n2]). The distance dL in cm between two figures in the interlocking image is estimated by the distance between two data points, (x[n1],y[n1]) and (x[n2],y[n2]), where n1 is the index at which the value of x[n1] is the maximum of those belonging to x[n] (nnf1) and n2 is the index at which the value of x[n2] is the minimum of those belonging to x[n] (nnf2). The ratio Δx/Δy is estimated by the ratio of the differences Δx and Δy, where Δx and Δy are the differences in cm in x- and y-axis directions, respectively, between two data points, (x[n1],y[n1]) and (x[n2],y[n2]). The existence and the shape of the interlocking between two figures can be discriminated by the percentage pL of the segmented interlocking image IL matching to the interlocking image IDL of the 128 × 128 drawing image ID, the distance dL, and the ratio Δx/Δy.

If the percentage pL is larger than a given threshold pth2, the distance dL has a negative value and the absolute value of the ratio |Δx/Δy| is larger than a given threshold rth, the two figures are then evaluated to be interlocked with a shape of a rhombus. If the percentage pL is larger than a threshold pth2 and the distance dL has a negative value but the absolute value of the ratio |Δx/Δy| is less than the given threshold rth, then the two figures are evaluated to be interlocked without the shape of a rhombus. On the other hand, if the distance dL has a positive value regardless of the percentage pL and the absolute value of the ratio |Δx/Δy|, then the two figures are evaluated as not to be interlocked and apart from each other by the distance dL. In such a case, if the value of the distance dL is nearly positively zero, being less than ε, a very small positive value, then the two figures are evaluated as to be attached to each other. The evaluation of the distance dL and the absolute ratio |Δx/Δy| are formulated as follows:

dL={(x[n1]x[n2])2+(y[n1]y[n2])2,x[n1]>x[n2](x[n1]x[n2])2+(y[n1]y[n2])2else,and
|Δx/Δy|=|x[n1]x[n2]y[n1]y[n1]|.

2.4.4. Existence of Closure/Opening

Existence of opening in a figure is determined by the fact that an opening could exist in a region with a relatively larger variance of distances between sample points, having consecutive 0s of velocity in a figure, as there are changes in the stroke direction as well as in the stroke position within the region where an opening occurs. Parameters of the existence of openings Nop1 and Nop2 in the figure images of ID,nnf1 and ID,nnf2 are estimated by k-means clustering of the sample points (x[n],y[n]), nnf1 and (x[n],y[n]), nnf2 with zero velocity, respectively, where the target numbers for the parameters Nop1 and Nop2 are set to the numbers of angles NA1 and NA2, respectively. Then, the cluster parameter kij (i=1,2,j=1,NAi) is calculated by averaging the distances between the sample points belonging to the cluster δij and the cluster center δc,ij as follows:

kij=mean(|(x[n],y[n])δc,ij|(x[n],y[n])δij).

The parameters of the existence of openings Nop1 and Nop2 in the figure images of ID,nnf1 and ID,nnf2 are then estimated as follows:

Nopi={1,(ki1>δth)oror(kiNAi>δth)0,else,

where δth is a given threshold.

2.4.5. Existence of Tremors

A tremor is an involuntary quivering movement or shake which could be caused by age-associated weakness, neurodegenerative diseases, or mental health conditions. A tremor can be detected from the frequency of a ‘down’ event happening after an ‘up’ event in the chain of touch-events, which cannot be observed in a paper and pencil test. The existence of tremors Ntri (i=1,2) in each figure is determined by the total number of ‘up’ events followed by ‘down’ events being larger than a given threshold eth in the touch-events e[n] of the sample points in the corresponding figure. An ‘up’ event followed by a ‘down’ event in the touch-events e[n] can be detected when the multiplication of two neighboring values in touch-events e[n] is equal to -1 as the values of ‘up’ and ‘down’ events are set to 1 and -1, respectively. The number of tremors Ntri (i=1,2) in each figure is evaluated as follows:

Ntr1={1,nnf1e[n]e[n+1]eth0,else,and
Ntr2={1,nnf2e[n]e[n+1eth0,else

.

2.4.6. Assignment of Scores

Table 3 lists the conditions for the assigned integer scores for each parameter in the mPDT. The score of the number of angles is via percentages, pf1 and pf2, in addition to the numbers of angles, NA1 and NA2. The score is a 4 if both of the percentages, pf1 and pf2, are equal or greater than a given threshold pth1; the score is a 3 if at least one of the percentages, pf1 or pf2, is less than pth1 as well as the sum number of angles, NA1+NA2, is 9 or 11; the score is a 2 if at least one of the percentages, pf1 or pf2, is less than pth1 as well as the sum number of angles, NA1+NA2, is 8 or 12; the score is a 1 if at least one of the percentages, pf1 or pf2, is less than pth1 as well as the sum number of angles, NA1+NA2, is between or including 5 and 7; and the score is a 0 if at least one of the percentages, pf1 or pf2, is less than pth1 as well as the sum number of angles, NA1+NA2, is less than 5 or greater than 12.

Table 3.

Details for the assignment of scores.

Parameters Assigned Integer Scores Conditions (Scoring Method)
Number of angles 4 pf1pth1 and pf2pth1
3 pf1<pth1 or pf2<pth1, NA1+NA2 = 9 or 11
2 pf1<pth1 or pf2<pth1, NA1+NA2 = 8 or 12
1 pf1<pth1 or pf2<pth1, 5 ≤ NA1+NA2 ≤ 7
0 pf1<pth1 or pf2<pth1, NA1+NA2 < 5 or >13
Distance/intersection 4 pLpth2, dL<ε, |Δx/Δy|rth
3 pL<pth2, dL<ε, |Δx/Δy|<rth
2 pL<pth2, |dL|ε
1 pL<pth2, ε<dL<1cm
0 pL<pth2, dL1cm
Closure/opening 2 Nop1Nop2=1, Nop1+Nop2=2
1 Nop1Nop2=0, Nop1+Nop2=1
0 Nop1Nop2=0, Nop1+Nop2=0
Tremor 1 Ntr1Ntr2=1
0 Ntr1Ntr2=0

The score for distance/intersection is obtained by using the percentage pL, the distance dL between the two figures in the interlocking image, and the absolute value of the ratio |Δx/Δy|. The score is a 4 if the percentage pL is equal to or larger than a given threshold pth2, the distance dL is less than ε, and the absolute value of the ratio |Δx/Δy| is equal to or larger than a given threshold rth; the score is a 3 if the percentage pL is less than pth2, the distance dL is less than ε, and the absolute value of the ratio |Δx/Δy| is less than a given threshold rth; the score is a 2 if the percentage pL is less than pth2 as well as the absolute of the distance dL is equal to and less than ε; the score is a 1 if the percentage pL is less than pth2 as well as the distance dL is between ε and 1 cm exclusive; and the score is a 0 if the percentage pL is less than pth2 as well as the distance dL is equal to or greater than 1 cm.

The score of closure/opening is determined by the parameters Nop1 and Nop2. The score is a 2 if there are no openings in both figures (Nop1Nop2 = 1 and Nop1+Nop2 = 2); the score is a 1 if there is an opening in one of the two figures (Nop1Nop2= 0 and Nop1+Nop2 = 1); and the score is a 0 if there are openings in both figures ( Nop1Nop2 = 0 and Nop1+Nop2 = 0).

The score for tremors is determined by the values Ntr1 and Ntr2. It is given a score of 1 if both Ntr1 and Ntr2 are equal to 1 (Ntr1Ntr2=1); and 0 if any of Ntr1 and Ntr2 values is equal to 0 (Ntr1Ntr2=0).

3. Results

3.1. Scoring of the Number of Angles

Figure 4 depicts separate examples of original drawings, each with their own characteristic shapes. The analytical ability of the Deep5 pre-trained model with its segmented images for a detected pentagon and the generated velocity graphs with the detected peaks for shape analysis are then demonstrated.

Figure 4.

Figure 4

Four examples of original drawings (left) along with their segmented images (middle left) produced by the pre-trained model, Deep5, their corresponding velocity graphs (middle right), and the corresponding parameter values (right). The number of angles was evaluated from the number of peaks detected in the velocity graph: (a) The image is composed of two interlocking figures, both being pentagons where the number of angles for both figures was evaluated to be 5 (pf1 and pf2 were 100.00% and 90.71%, respectively, and both greater than 75%). (b) The image of two interlocking figures, but with only one being a pentagon (the left-side figure). The number of angles of the left-side figure was evaluated to be 5, as pf1 was 91.5% and it was greater than 75%. For the right-side figure, the number of angles was given as 4, as the percentage pf2 of 23.88% was less than 75% and the number of peaks detected in the velocity graph is 4. (c) The image of the right portion is a pentagon with the number of angles evaluated as 5 as pf2 was 98.52% and it was greater than 75%. The image of the left portion was evaluated as a 4, as pf1 was 29.10% and it was less than 75% and the number of peaks detected in the velocity graph is 4. (d) The image, composite of two figures, none of them being a pentagon. This is because the number of angles for each of the right and left-hand portions was evaluated to be a 4, since pf1 and pf2 were 18.18% and 0.00%, respectively, both being less than 75%. The numbers of angles in both the right and left portions were evaluated to be 4, matching the number of peaks detected in the velocity graphs for both.

In Figure 4a where the original image is of two interlocking figures (left), both being pentagons, the segmented image (center) perceived by Deep5 has the estimated percentages pf1 and pf2 of 100.00% and 90.71%, respectively. The number of angles in each pentagon were both evaluated to be 5, as pf1 and pf2 were greater than the 0.75 score for pth1, a threshold heuristically set by the two clinical experts in PDT scaling during the process of the ground truth scorings of all the images used in this study. For the velocity graph, the number of detected peaks corresponding to the number of angles in each figure was evaluated to be 5 (right). Figure 4b has the original drawing image of two interlocking figures, but only one is a pentagon (left figure). The segmented image by Deep5 has the estimated percentages pf1 and pf2 of 91.5% and 23.88%, respectively (center). The number of angles in the pentagon portion was evaluated to be 5, as the estimated percentage pf1 was greater than 0.75; however, the number of angles in the non-pentagon portion was gauged to be 4, as the estimated percentage pf2 was less than the 0.75 score and the number of peaks detected in the velocity graph was estimated to be 4 as well (right). Figure 4c similarly shows an example of an original drawing (left) of two interlocking figures with only one being a pentagon. The segmented image (middle) by Deep5 where the estimated percentages pf1 and pf2 were 29.10% and 98.52%, respectively. The number of angles in the non-pentagon portion was evaluated to be 4 from the estimated percentage score, pf1, being less than 0.75. The number of peaks in the non-pentagon portion of the figure from the velocity graph (right) was estimated to be 4 as well. On the other hand, the number of angles of the pentagon portion was evaluated to be 5, as the estimated percentage score, pf2, was greater than 0.75. Finally, the original drawing image (left) of Figure 4d depicts the example of two interlocking figures, with none of them being a pentagon. For the segmented image (middle) by Deep5, the estimated percentages, pf1 and pf2 were estimated to be 18.18 and 0.00%, respectively. In this case, the number of angles of both non-pentagons was evaluated to be 4, as both estimated percentages were less than 0.75 and the numbers of peaks detected in the velocity graph (right) were also estimated to be 4 for both non-pentagons.

3.2. Scoreing of Distance/Interlocking

The analytical ability of the pre-trained model DeepLock is demonstrated in Figure 5 with five separate examples of original drawing image along with the segmented image of interlocking generated by the program. Figure 5a has the original drawing image example (left) with the interlocking shape of a rhombus, its segmented image (middle) interlocking, and the overlap of the original drawing image along with the segmented image (right). In this case, the percentage pL and the distance dL were evaluated to be 97.67% and −2.27 cm, respectively. The absolute value |Δx/Δy| of the application was estimated to be 4.49. As such, the two figures were evaluated to be interlocked with a shape of a rhombus from three parameters: (1) percentage score, pL, being greater than pth2 of 0.75, a heuristically given threshold; (2) the absolute value |Δx/Δy| being larger than 1.12, a heuristically given threshold of rth, and (3) the distance dL of a negative value equal to and less than 0.01 cm, as the threshold value of ε. Here, the threshold values for pth2 and rth were chosen by the two clinical experts in PDT scaling during the process of the ground truth scorings of all the images used in this study. The threshold value ε was set to be 0.01 cm, considering a 2 pixel diagonal distance, 0.004 × 2 × sqrt(2), with a spatial resolution 0.004 cm of the device used in the implementation of the mPDT. The example in Figure 5b has the original drawing image of an interlocking shape that is not a rhombus (left) along with the resulting segmented image (middle) of the interlocking, and the overlap of the original drawing image and the segmented image (right). In this case, the percentage pL was evaluated to be 99.79% and the distance dL to be -2.36 cm. The absolute value |Δx/Δy| for the plot was estimated to be 0.86. From these parameters, the two figures were evaluated to be interlocked without a shape of a rhombus, having the percentage pL of greater than 0.75 for pth2 and the distance dL of a negative value equal to and less than 0.01 cm for ε, but the absolute value |Δx/Δy| being less than 1.12 for rth.

Figure 5.

Figure 5

Five examples of the original drawing data along with the segmented images generated and analyzed by the pre-trained model, DeepLock (left and/or middle) and the corresponding parameter values (right). Below, the values in parentheses are the set comparator values for that parameter: (a) The case of the two interlocking figures giving a shape of a rhombus, where the percentage pL, the distance dL and the absolute value |Δx/Δy| of the ratio were evaluated to be 97.67% ( > 0.75%), −2.27 cm ( < −0.01 cm) and 4.49 ( > 1.12), respectively. (b) The example of two interlocking figures, not giving a shape of a rhombus, where the percentage pL, the distance dL and the absolute value |Δx/Δy| of the ratio were evaluated to be 99.79% ( > 0.75%), −2.36 cm (< −0.01 cm) and 0.86 ( < 1.12), respectively. (c) The case of two figures with no interlocking but still touching each other. There, the percentage pL and the distance dL were evaluated to be 0.00% ( < 0.75%) and 0.001 cm (being in between −0.01 and 0.01 cm range), respectively. (d) The display case of two figures, drawn so that they are separated from each other, but are within 1.00 cm of each other. The percentage pL and the distance dL were evaluated to be 0.00% ( < 0.75%) and 0.24 cm, respectively. (e) The example of two drawn figures that are more than 1.00 cm apart with the percentage pL and the distance dL were evaluated to be 0.00% (< 0.75%) and 1.50 cm, respectively.

In the example for Figure 5c, the two figures are not intersecting, but are making contact. The percentage pL was estimated to be 0.00% and the distance dL was calculated to be 0.001 cm, less than 0.01 cm for ε. In Figure 5d, the original image depicts two component figures that are apart from each other; the percentage pL was estimated to be 0.00% and the distance dL was calculated as to be 0.24 cm, greater than ε of 0.01 cm. The example in Figure 5e has two figures that are also apart from each other, giving the percentage pL estimated as 0.00% and the distance dL calculated as 1.50 cm.

3.3. Scoring of Closure/Opening

In three representative examples, Figure 6 demonstrates how the openings in original images are detected and assigned. Figure 6a displays the case of an original image having no openings in the two interlockings. For the two interlocking figures, the cluster parameters k1j (j=1,NA1) of the left interlocking figure were 0.014, 0.049, 0.037, 0.014, and 0.024 cm, respectively; for the right interlocking figure, k2j (j=1,NA2) were 0.060, 0.014, 0.012, and 0.063 cm, respectively. Both sets of k1j and k2j values were less than the threshold δth of 0.1 cm for all their individual values. For Figure 6b, of the two interlocking figures, there is an opening for the left figure. k1j (j=1,NA1) values for the left figure were 0.054, 0.080, 0.015, 0.046, and 0.179 cm, respectively, with one being greater than δth of 0.1 cm. For the right portion of the figure, the k2j (j=1,NA2) values were 0.027, 0.014, 0.022, 0.011, and 0.035 cm, respectively, all of less than δth of 0.1 cm. The example in Figure 6c has an opening in each of the two interlocking figures. k1j (j=1,NA1) for the left figure were 0.015, 0.019, 0.023, 0.023, and 0.193 cm, respectively, one of which was greater than δth of 0.1 cm. The k2j (j=1,NA2) values for the right figure were 0.014, 0.013, 0.014, 0.011, and 0.167 cm, respectively, again with one being greater than δth of 0.1 cm, demonstrating an opening in the figure. Here, the threshold value δth was set to be 0.1 cm considering the line width (set to 20 pixels, ~0.08 cm) and the spatial resolution 0.004 cm of the device used in the implementation of the mPDT.

Figure 6.

Figure 6

Three examples of openings being detected: (a) The case without any openings in both drawn figures. Here, the cluster parameters k1j and k2j of the left and right-portion figures were 0.014, 0.049, 0.037, 0.014, and 0.024 cm and 0.060, 0.014, 0.012, and 0.063 cm, respectively (all < 0.1 cm). (b) The example of a drawing with openings only in the left-portion of the figure. The cluster parameters k1j for the left-portion figure were 0.054, 0.080, 0.015, 0.046, and 0.179 cm (thus one >0.1cm); however, the cluster parameters k2j for the right-portion of the figure were 0.027, 0.014, 0.022, 0.011, and 0.035 cm (all < 0.1 cm). (c) The case of openings in both drawn figures with the cluster parameters k1j and k2j of the left and right figures of 0.015, 0.019, 0.023, 0.023, and 0.193 cm and 0.014, 0.013, 0.014, 0.011, and 0.167 cm, respectively (one for each figure > 0.1 cm).

3.4. Scoring of Tremors

For the presence of tremors in the hand drawn images, Figure 7 represents two cases with one having no tremors and the other having tremors. For Figure 7a with no significant tremors being present in either of the two interlocking figures, the total numbers of the ‘up’ events followed by ‘down’ events in touch-events e[n] of the sample points for the left and right portions of the interlocking figure were 0 and 1, respectively, and both with a given threshold eth of less than 5. The threshold value eth here was set by the two clinical experts in PDT scaling during the process of the ground truth scorings of all the images used in this study. In contrast, Figure 7b displays drawings with some tremors present in each of the two interlocking figures. In the figure, the ‘up’ events followed by ‘down’ events in the touch-events e[n] were 5 and 19 for the left and right figures, respectively, both being equal to or greater than 5, the given threshold eth.

Figure 7.

Figure 7

Two examples for detection of tremors: (a) The case without a detectable tremor, where the numbers of the ‘up’ events followed by ‘down’ events in the recorded touch-events were 0 and 1 for the left and right-portion figures, respectively (both < 5). (b) The case with detectable tremors where the numbers of the ‘up’ events followed by ‘down’ events in the recorded touch-events were 5 and 19 for the left and right-portion figures, respectively (one was >5).

3.5. Performance Test Results

A total of 230 drawing images were used to test the performance of the scoring method with the mPDT. Table 4 summarizes the frequency of the ground truth for the 230 images with the score in each of the parameters. For the number of angles detected, the scores of 0 through 4 were 55, 32, 33, 30, and 80 events. For the distance/intersection parameter, the scores of 0 through 4 were for 33, 38, 34, 38, and 87 measures. Similarly, for the closure/opening measure, the scores of 0 through 2 were for 39, 61, and 130 detections, in the given order. The total numbers of instances for absence or presence of tremors, with a score of 0 or 1, were 16 and 214, respectively.

Table 4.

Frequency of the ground truth of the 230 images by score in each parameter of the scoring method of the mPDT.

Scores Number of Angles Distance/Intersection Closure/Opening Tremor
4 80 87 - -
3 30 38 - -
2 33 34 130 -
1 32 38 61 214
0 55 33 39 16
total 230 230 230 230

Table 5 lists the performance of each scoring parameter in mPDT. For the angle number parameter, the sensitivity, specificity, accuracy, and precision values were 97.53%, 92.62%, 94.35%, and 87.78%; for distance/intersection, they were 93.10%, 97.90%, 96.09%, and 96.43%; for closure/opening, they were 94.03%, 90.63%, 92.61%, and 93.33%; and for tremor reads, they were 100.00%, 100.00%, 100.00%, and 100.00%, respectively.

Table 5.

Performance of the scoring parameters in the mPDT.

Number of Angles Distance/Intersection Closure/Opening Tremor
TP 79 81 126 214
FP 11 3 9 0
FN 2 6 8 0
TN 138 140 87 16
Sensitivity 97.53 93.10 94.03 100.00
Specificity 92.62 97.90 90.63 100.00
Accuracy 94.35 96.09 92.61 100.00
Precision 87.78 96.43 93.33 100.00

4. Discussion

Conventional PDT based on a paper and pencil test is not readily suitable for evaluation of the dynamic components of cognitive function, as there are limitations in the real-time tracking of the orders, the stroke patterns, the speed variations, and so on, while the subjects are copying or drawing from recall. When subjects participate in a conventional PDT, many fMRI studies have shown multiple brain areas becoming active in the subject, including the bilateral parietal lobe, sensorimotor cortex, cerebellum, thalamus, premotor area, and inferior temporal sulcus [33,34,35,36]. However, it is not exactly clear what components of the cognitive function are associated with the activation of these areas as the conventional PDT is difficult to quantify objectively. To address this issue, our study focused on implementation of the PDT as a mobile phone application, namely mPDT, with a novel, automatic, and qualitative scoring method based on U-Net, a convolutional network for biomedical image segmentation of sensor data. The sensor data is also obtained by the mPDT.

The performance test proved that the scoring protocol suggested by the mPDT is reasonable and practical when compared with those of the traditional PDT. Further, the mPDT was shown to be capable of evaluation of the dynamic components of cognitive function. In our study, the subjects used a smartpen provided for the smartphone when copying figures in order to create an environment similar to the conventional paper and pencil test of the PDT. This also increased the accuracy and avoided undesirable noise in the activated brain function assay. The performance test was restricted to right-handed subjects to avoid a bias in statistical analysis and also due to relatively small number of left-handed subjects available. However, the mPDT scoring worked quite in the same way when two left-handed subjects (a 27 year old male and a 26 year old female) were initially included in the younger volunteer group, both samples showing proper results and accuracies (see Appendix C).

The conventional PDT is a sub-item of MMSE, which is usually used in assessing Alzheimer’s disease [25]. However, the mPDT was developed to be applicable in better detection of cognitive impairment in Parkinson’s disease. For this reason, the tremor parameter was included in the scoring of the mPDT, instead of the closing-in parameter suggested in previous qualitative scoring of the pentagon test [5], and as the closing-in sign is a characteristic sign in all dementia, but not for Parkinson’s disease [37]. A tremor is an involuntary quivering movement or shake which may be due to age-associated weakness, a neurodegenerative disease, or a mental health condition. Therefore, there are several types of tremors recognized, such as essential tremors, Parkinsonian tremors, dystonic tremors, cerebellum tremors, psychogenic tremors, orthostatic tremors, and physiologic tremors [38]. In this study, the tremor symptom could be detected by the frequency of the occurrence of a ‘down’ event after an ‘up’ event in the touch-event series. A future study using mPDT for correlations between the pattern of tremoring and the underlying disease or condition could make a case for an early and differential diagnosis of a given neurodegenerative disease, such as Parkinson’s.

Conventional screening tools, including MMSE, do not detect early cognitive impairment in Parkinson’s disease, while the PDT is known to detect cognitive impairment earlier in Parkinson’s than in Alzheimer’s disease [39]. Using this fact, we aimed to develop a more sensitive screening tool to detect cognitive impairment in Parkinson’s. In addition, as the smartphone could evaluate the motor-related indicators such as the speed at which the test was performed or the number of pauses, we could also measure the effects on motor ability as a cognitive measurement tool that could not be detected by the conventional PDT pencil and paper test. Therefore, our study indicates that the developed mPDT tool is specifically applicable in increasing the accuracy of cognitive function assessment in Parkinson’s disease.

5. Conclusions

Even though the qualitative scorings of the PDT have been essential in establishing it as a prognostic marker in the assessment of visuospatial functions and in the differentiation of various neuronal degenerative diseases including Parkinson’s, the evaluation is done manually, which is not subjective, is prone to human error, and it is not able to provide parametric and dynamic information on a specific neuronal degenerative disease. In this study, we developed a smartphone application, named mPDT, with an automatic scoring method based on mobile sensors and image segmentation using the U-Net deep learning algorithm. A tremor read, not in the standard PDT, was also included, allowing for the detection of early Parkinson’s along with the other parameters tested.

The mPDT is also relatively environment independent as it is applicable for different types of mobile devices, such as smartphones, tablets, and notebooks. It is also relatively fast. The execution time was 0.73 ± 0.21 seconds (mean ± std.) for the total score after a drawn PDT image was submitted in the performance test for a machine with Processor Intel(R) Core(TM) i7-8700 CPU at 3.20 GHz, 3192 Mhz 6 Core(s), 12 Logical Processor(s) with 8 GB of RAM running 64-bit Windows Version 10 specifications.

The mPDT is very easy, simple and intuitive to use and it can be convenient for use by the elderly. Using mPDT for the PDT test also allows evaluation of the results objectively and qualitatively as well as for parametric assessment of the results. This can also allow differentiation of the dynamic components of the cognitive function underlying a given neurodegenerative disease. Furthermore, because redrawing and saving of the sensor data along with the images drawn by subjects are possible in any mobile, electronic device, the onset and time course of brain neuronal degeneration could be detected and monitored as a basis of a personal lifelog as well as in real time. Therefore, this tool is to evaluate the current cognitive functions of the examinee and better distinguish the causes of cognitive decline.

For future work, we are currently developing qualitative and automatic scoring algorithms for the CDT and the ROCF tests by expanding the algorithms used in mPDT. Directions for this future work include various drawing tests, such as a draw-a-person test, a draw-a-family test, and so on, which would need more specific deep neural networks for image segmentation, feature extraction and classification, and clustering correlations between features.

Appendix A

Table A1.

Summary and comparison of the present study and the existing literature.

Mode of Drawing Test Type Scoring Method and Spec. Reference
Clock drawing test (CDT) paper-based quantitative, semi-quantitative
5, 6, 10, 12, 20 points systems
manually interpreting
[16]
2018
review
digital CDT semi-quantitative
automatic estimation of stroke features, up and down
6 points system
manually interpreting based on computerized feature
[17]
2017
digital CDT qualitative
ontology-based knowledge representation
CNN for object recognition
automatically recognize each number by the probability score
[18]
2017
digital CDT qualitative
categorized stroke data based on feature using machine learning
[19]
2016
digital CDT qualitative
automatic estimation of stroke features, up and down, stroke velocity, stroke pressure
[20]
2019
Rey–Osterrieth complex figure (ROCF) paper-based quantitative
location and perceptual grading of the basic geometric features
[21]
1944
paper-based qualitative
points 0-24 based on the order in which the figure is produced
[23]
2017
paper-based quantitative
automated scoring of 18 segments based on a cascade of deep neural networks trained on human rater scores
[24]
2019
PDT paper-based qualitative
6 points system
manually interpreting
[25]
1995
paper-based qualitative
10 points system
manually interpreting
[26]
2011
paper-based qualitative
6 sub scales
manually interpreting based on factor analysis and 6 subscales correlated to control responses
[27]
2012
paper-based qualitative
points 0–13 based on parametric estimations for number of angles, distance/intersection, closure/opening, rotation, closing-in
manually interpreting
[5]
2013
mPDT mobile-based qualitative
points 0–11 based on stroke feature for the order and the speed and also parametric estimations for number of angles, distance/intersection, closure/opening, tremor
automatic interpreting using U-net deep learning algorithm and sensor data
The presented system

Appendix B

Table A2.

Examples of corresponding drawings and scores for each level with the corresponding detailed parameters.

Case Image Performance Scores Assigned Integer Scores Total Score
NA1 D/I2 C/O3 Tr4 NA1 D/I2 C/O3 Tr4
graphic file with name sensors-20-01283-i001.jpg 10 CI5 NO10 NT13 4 4 2 1 11
graphic file with name sensors-20-01283-i002.jpg 10 CI5 O111 NT13 4 4 1 1 10
graphic file with name sensors-20-01283-i003.jpg 10 WI6 NO10 NT13 4 3 2 1 10
graphic file with name sensors-20-01283-i004.jpg 10 CI5 O212 NT13 4 4 0 1 9
graphic file with name sensors-20-01283-i005.jpg 10 WI6 O111 NT13 4 3 1 1 9
graphic file with name sensors-20-01283-i006.jpg 10 NC08 NO10 NT13 4 1 2 1 8
graphic file with name sensors-20-01283-i007.jpg 10 C7 O111 NT13 4 2 1 1 8
graphic file with name sensors-20-01283-i008.jpg 9 CI5 NO10 NT13 3 4 2 1 10
graphic file with name sensors-20-01283-i009.jpg 9 CI5 O111 NT13 3 4 1 1 9
graphic file with name sensors-20-01283-i010.jpg 9 WI6 NO10 NT13 3 3 2 1 9
graphic file with name sensors-20-01283-i011.jpg 9 C7 NO10 NT13 3 2 2 1 8
graphic file with name sensors-20-01283-i012.jpg 9 WI6 O111 NT13 3 3 1 1 8
graphic file with name sensors-20-01283-i013.jpg 9 WI6 O212 NT13 3 3 0 1 7
graphic file with name sensors-20-01283-i014.jpg 9 C7 O111 NT13 3 2 1 1 7
graphic file with name sensors-20-01283-i015.jpg 11 NC08 O212 NT13 3 1 0 1 5
graphic file with name sensors-20-01283-i016.jpg 8 CI5 NO10 NT13 2 4 2 1 9
graphic file with name sensors-20-01283-i017.jpg 8 CI5 NO10 NT13 2 4 2 1 9
graphic file with name sensors-20-01283-i018.jpg 8 WI6 NO10 NT13 2 3 2 1 8
graphic file with name sensors-20-01283-i019.jpg 8 WI6 NO10 NT13 2 3 2 1 8
graphic file with name sensors-20-01283-i020.jpg 8 WI6 O111 NT13 2 3 1 1 7
graphic file with name sensors-20-01283-i021.jpg 8 C7 NO10 NT13 2 2 2 1 7
graphic file with name sensors-20-01283-i022.jpg 8 NC08 NO10 NT13 2 1 2 1 6
graphic file with name sensors-20-01283-i023.jpg 8 NC08 O111 NT13 2 1 1 1 5
graphic file with name sensors-20-01283-i024.jpg 8 NC19 NO10 NT13 2 0 2 1 5
graphic file with name sensors-20-01283-i025.jpg 8 NC08 O212 NT13 2 1 0 1 4
graphic file with name sensors-20-01283-i026.jpg 7 CI5 NO10 NT13 1 4 2 1 8
graphic file with name sensors-20-01283-i027.jpg 7 NC08 NO10 NT13 1 1 2 1 5
graphic file with name sensors-20-01283-i028.jpg 7 NC19 NO10 NT13 1 0 2 1 4
graphic file with name sensors-20-01283-i029.jpg 7 NC19 O111 NT13 1 0 1 1 3
graphic file with name sensors-20-01283-i030.jpg 6 NC19 O111 NT13 1 0 1 1 3
graphic file with name sensors-20-01283-i031.jpg >13 WI6 NO10 NT13 0 3 2 1 6
graphic file with name sensors-20-01283-i032.jpg >13 CI5 O212 NT13 0 4 0 1 5
graphic file with name sensors-20-01283-i033.jpg >13 NC08 O111 NT13 0 1 1 0 2
graphic file with name sensors-20-01283-i034.jpg >13 C7 O212 T14 0 2 0 0 2
graphic file with name sensors-20-01283-i035.jpg 4 NC19 O212 NT13 0 0 0 1 1
graphic file with name sensors-20-01283-i036.jpg 4 NC19 O212 T14 0 0 0 0 0

1NA: Number of angles; 2D/I: Distance/interaction; 3C/O: Closure/opening; 4Tr: Tremor. 5CI: Correct intersection; 6WI: Wrong intersection; 7C: Contact without intersection; 8NC0: No contact, distance<1cm; 9NC1: No contact, distance>1cm; 10NO: Closure both figures; 11O1: Closure only one figure; 12O2: Opening both figures; 13NT: No Tremor; 14T: Tremors.

Appendix C

Figure A1.

Figure A1

A case of a left-handed male subject (27 years old): (a) the original drawing (left) along with the segmented image (middle left) produced by the pre-trained model, Deep5, the corresponding velocity graph (middle right), the corresponding parameter values (right); (b) the segmented images generated and analyzed by the pre-trained model, DeepLock (left and middle) and the corresponding parameter values (right); (c) the clusters marked on the original drawing (left) and the corresponding parameter values (right); and (d) the up and down points marked on the original drawing (left) as well as on the velocity graph (middle) and the corresponding parameter values (right).

Figure A2.

Figure A2

A case of a left-handed female subject (26 years old): (a) the original drawing (left) along with the segmented image (middle left) produced by the pre-trained model, Deep5, the corresponding velocity graph (middle right), the corresponding parameter values (right); (b) the segmented images generated and analyzed by the pre-trained model, DeepLock (left and middle) and the corresponding parameter values (right); (c) the clusters marked on the original drawing (left) and the corresponding parameter values (right); and (d) the up and down points marked on the original drawing (left) as well as on the velocity graph (middle) and the corresponding parameter values (right).

Table A3.

The corresponding drawings and scores for each level with the corresponding detailed parameters of the case of two left-handed subjects.

Case of Image of the Left-Handed Subject Performance Scores Assigned Integer Scores Total Score
NA1 D/I2 C/O3 Tr4 NA1 D/I2 C/O3 Tr4
graphic file with name sensors-20-01283-i037.jpg 10 CI5 NO6 NT7 4 4 2 1 11
graphic file with name sensors-20-01283-i038.jpg 10 CI5 NO6 NT7 4 4 2 1 11

1NA: Number of angles; 2D/I: Distance/interaction; 3C/O: Closure/opening; 4Tr: Tremor; 5CI: Correct intersection; 6NO: Closure both figures; 7NT: No Tremors.

Author Contributions

Conceptualization, U.L. and Y.J.K. (Yeo Jin Kim); methodology, U.L.; software development, U.L. and I.P.; validation, U.L., Y.J.K. (Yun Joong Kim) and Y.J.K. (Yeo Jin Kim); formal analysis, U.L.; investigation, U.L. and Y.J.K. (Yun Joong Kim); resources, U.L., I.P., Y.J.K. (Yun Joong Kim) and Y.J.K. (Yeo Jin Kim); data curation, U.L. and I.P.; writing—original draft preparation, U.L. and Y.J.K. (Yeo Jin Kim); writing—review and editing, U.L. and Y.J.K. (Yeo Jin Kim); visualization, U.L. and I.P.; supervision, U.L. and Y.J.K. (Yeo Jin Kim); project administration, U.L. and Y.J.K. (Yeo Jin Kim); funding acquisition, U.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Brain Research Program (2017M3A9F1030063) of the National Research Foundation of Korea and Hallym University Research Fund (HRF-201902-015).

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Folstein M.F., Folstein S.E., McHugh P.R. Mini-mental state: A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 1975;12:189–198. doi: 10.1016/0022-3956(75)90026-6. [DOI] [PubMed] [Google Scholar]
  • 2.Biundo R., Weis L., Bostantjopoulou S., Stefanova E., Falup-Pecurariu C., Kramberger M.G., Geurtsen G.J., Antonini A., Weintraub D., Aarsland D. MMSE and MoCA in Parkinson’s disease and dementia with Lewy bodies: A multicenter 1-year follow-up study. J. Neural Transm. 2016;123:431–438. doi: 10.1007/s00702-016-1517-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Błaszczyk J.W. Parkinson’s Disease and Neurodegeneration: GABA-Collapse Hypothesis. Front. Neurosci. 2016;10 doi: 10.3389/fnins.2016.00269. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Hely M.A., Reid W.G., Adena M.A., Halliday G.M., Morris J.G. The Sydney Multicenter Study of Parkinson’s Disease: The Inevitability of Dementia at 20 years. Mov. Disord. 2008;23:837–844. doi: 10.1002/mds.21956. [DOI] [PubMed] [Google Scholar]
  • 5.Caffarraa P., Gardinia S., Diecib F., Copellib S., Maseta L., Concaria L., Farinac E., Grossid E. The qualitative scoring MMSE pentagon test (QSPT): A new method for differentiating dementia with Lewy Body from Alzheimer’s Disease. Behav. Neurol. 2013;27:213–220. doi: 10.1155/2013/728158. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Cecato J.F. Pentagon Drawing Test: Some data from Alzheimer’s disease, Paraphrenia and Obsessive compulsive disorder in elderly patients. Perspect. Psicol. 2016;13:21–26. [Google Scholar]
  • 7.Ala T.A., Hughes L.F., Kyrouac G.A., Ghobrial M.W., Elble R.J. Pentagon copying is more impaired in dementia with Lewy bodies than in Alzheimer’s disease. J. Neurol. Neurosurg. Psychiatry. 2001;70:483–488. doi: 10.1136/jnnp.70.4.483. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Teng E.L., Chui H.C. The modified Mini-Mental State (3-MS) Examination. J. Clin. Psychiatry. 1987;48:314–318. [PubMed] [Google Scholar]
  • 9.Martinelli J.E., Cecato J.F., Msrtinelli M.O., Melo B.A.R., Aprahamian I. Performance of the Pentagon Drawing test for the screening of older adults with Alzheimer’s dementia. Dement. Neuropsychol. 2018;12:54–60. doi: 10.1590/1980-57642018dn12-010008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Lam M.W.Y., Liu X., Meng H.M.L., Tsoi K.K.F. Drawing-Based Automatic Dementia Screening Using Gaussian Process Markov Chains; Proceedings of the 51st Hawaii International Conference on System Sciences; Hilton Waikoloa Village, HI, USA. 1 January 2018. [Google Scholar]
  • 11.Mervis C.B., Robinson B.F., Pani J.R. Cognitive and behavioral genetics ’99 visuospatial construction. Am. J. Hum. Genet. 1999;65:1222–1229. doi: 10.1086/302633. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Johnson D.K., Storandt M., Morris J.C., Galvin J.E. Longitudinal study of the transition from healthy aging to Alzheimer disease. Arch. Neurol. 2009;66:1254–1259. doi: 10.1001/archneurol.2009.158. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Aprahamian I., Martinelli J.E., Neri A.L., Yassuda M.S. The clock drawing test: A review of its accuracy in screening for dementia. Dement. Neuropsychol. 2009;3:74–80. doi: 10.1590/S1980-57642009DN30200002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.McGee S. Evidence-Based Physical Diagnosis. 4th ed. Elsevier; Amsterdam, The Netherlands: 2018. pp. 49–56. [Google Scholar]
  • 15.Spenciere B., Alves H., Charchat-Fichman H. Scoring systems for the Clock Drawing Test, A historical review. Dement. Neuropsychol. 2017;11:6–14. doi: 10.1590/1980-57642016dn11-010003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Palsetia D., Prasad Rao G., Tiwari S.C., Lodha P., De Sousa A. The Clock Drawing Test versus Mini-mental Status Examination as a Screening Tool for Dementia: A Clinical Comparison. Indian J. Psychol. Med. 2018;40:1–10. doi: 10.4103/IJPSYM.IJPSYM_244_17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Müller S., Preische O., Heymann P., Elbing U., Laske C. Increased Diagnostic Accuracy of Digital vs. Conventional Clock Drawing Test for Discrimination of Patients in the Early Course of Alzheimer’s Disease from Cognitively Healthy Individuals. Front. Aging Neurosci. 2017;9 doi: 10.3389/fnagi.2017.00101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Harbi Z., Hicks Y., Setchi R. Clock Drawing Test Interpretation System; Proceedings of the International Conference on Knowledge Based and Intelligent Information and Engineering Systems, KES2017; Marseille, France. September 2017. [Google Scholar]
  • 19.Souillard-Mandar W., Davis R., Rudin C., Au R., Libon D.J., Swenson R., Price C.C., Lamar M., Penney D.L. Learning Classification Models of Cognitive Conditions from Subtle Behaviors in the Digital Clock Drawing Test. Mach. Learn. 2016;102:393–441. doi: 10.1007/s10994-015-5529-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Müller S., Herde L., Preische O., Zeller A., Heymann P., Robens S., Elbing U., Laske C. Diagnostic value of digital clock drawing test in comparison with CERAD neuropsychological battery total score for discrimination of patients in the early course of Alzheimer’s disease from healthy Individuals. Sci. Rep. 2019;9:3543. doi: 10.1038/s41598-019-40010-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Osterrieth P.A. Filetest de copie d’une figure complex: Contribution a l’etude de la perception et de la memoire The test of copying a complex figure: A contribution to the study of perception and memory. Arch. Psychol. 1944;30:286–356. [Google Scholar]
  • 22.Canham R.O., Smith S.L., Tyrrell A.M. Automated Scoring of a Neuropsychological Test: The Rey Osterrieth Complex Figure; Proceedings of the 26th Euromicro Conference. EUROMICRO 2000. Informatics: Inventing the Future; Maastricht, The Netherlands. 5–7 September 2000; pp. 406–413. [Google Scholar]
  • 23.Sargenius H.L., Bylsma F.W., Lydersen S., Hestad K. Visual-Constructional Ability in Individuals with Severe Obesity: Rey Complex Figure Test Accuracy and the Q-Score. Front. Psychol. 2017;8:1629. doi: 10.3389/fpsyg.2017.01629. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Vogt J., Kloosterman H., Vermeent S., Van Elswijk G., Dotsch R., Schmand B. Automated scoring of the Rey-Osterrieth Complex Figure Test using a deep-learning algorithm. Arch. Clin. Neuropsychol. 2019;34:836. doi: 10.1093/arclin/acz035.04. [DOI] [Google Scholar]
  • 25.Bourke J., Castleden C.M., Stephen R., Dennis M. A comparison of clock and pentagon drawing in Alzheimer’s disease. Int. J. Geriatr. Psychiatry. 1995;10:703–705. doi: 10.1002/gps.930100811. [DOI] [Google Scholar]
  • 26.Nagaratnam N., Nagaratnam K., O’Mara D. Intersecting pentagon copying and clock drawing test in mild and moderate Alzheimer’s disease. J. Clin. Gerontol. Geriatr. 2014;5:47–52. doi: 10.1016/j.jcgg.2013.11.001. [DOI] [Google Scholar]
  • 27.Fountoulakis K.N., Siamouli M., Panagiotidis P.T., Magiria S., Kantartzis S., Terzoglou V.A., Oral T. The standardised copy of pentagons test. Ann. Gen. Psychiatry. 2011 doi: 10.1186/1744-859X-10-13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Lee E.-S., Hong Y.J., Yoon B., Lim S.-C., Shim Y.S., Ahn K.J., Cho A.-H., Yang D.-W. Anatomical Correlates of Interlocking Pentagon Drawing. Dement. Neurocogn. Disord. 2012;11:141–145. doi: 10.12779/dnd.2012.11.4.141. [DOI] [Google Scholar]
  • 29.Cormack F., Aarsland D., Ballard C., Tovée M.J. Pentagon drawing and neuropsychological performance in Dementia with Lewy Bodies, Alzheimer’s disease, Parkinson’s disease and Parkinson’s disease with dementia. Int. J. Geriatr. Psychiatry. 2004;19:371–377. doi: 10.1002/gps.1094. [DOI] [PubMed] [Google Scholar]
  • 30.Beretta L., Caminiti S.P., Santangelo R., Magnani G., Ferrari-Pellegrini F., Caffarra P., Perani D. Two distinct pathological substrates associated with MMSE-pentagons item deficit in DLB and AD. Neuropsychologia. 2019;133:104–174. doi: 10.1016/j.neuropsychologia.2019.107174. [DOI] [PubMed] [Google Scholar]
  • 31.Vergouw L.J.M., Salomé M., Kerklaan A.G., Kies C., Roks G., van den Berg E., de Jong F.J. The Pentagon Copying Test and the Clock Drawing Test as Prognostic Markers in Dementia with Lewy Bodies. Dement. Geriatr. Cogn. Disord. 2018;45:308–317. doi: 10.1159/000490045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Ronneberger O., Fischer P., Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N., Hornegger J., Wells W., Frangi A., editors. Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015. Springer; Cham, Switzerland: 2015. [Google Scholar]
  • 33.Matt E., Foki T., Fischmeister F., Pirker W., Haubenberger D., Rath J., Lehrner J., Auff E., Beisteiner R. Early dysfunctions of fronto-parietal praxis networks in Parkinson’s disease. Brain Imaging Behav. 2017;11:512–525. doi: 10.1007/s11682-016-9532-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Potgieser A.R.E., van der Hoorn A., de Jong B.M. Cerebral Activations Related to Writing and Drawing with Each Hand. PLoS ONE. 2015 doi: 10.1371/journal.pone.0126723. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Segal E., Petrides M. The anterior superior parietal lobule and its interactions with language and motor areas during writing. Eur. J. Neurosci. 2012;35:309–322. doi: 10.1111/j.1460-9568.2011.07937.x. [DOI] [PubMed] [Google Scholar]
  • 36.Makuuchi M., Kaminaga T., Sugishita M. Both parietal lobes are involved in drawing: A functional MRI study and implications for constructional apraxia. Cogn. Brain Res. 2003;16:338–347. doi: 10.1016/S0926-6410(02)00302-6. [DOI] [PubMed] [Google Scholar]
  • 37.De Lucia N., Grossi D., Trojano L. The genesis of closing-in in alzheimer disease and vascular dementia: A comparative clinical and experimental study. Neurophysiology. 2014;28:312–318. doi: 10.1037/neu0000036. [DOI] [PubMed] [Google Scholar]
  • 38.Ryu H.-S. New Classification of Tremors: 2018 International Parkinson and Movement Disorder Society. J. Korean Neurol. Assoc. 2019;37:251–261. doi: 10.17340/jkna.2019.3.2. [DOI] [Google Scholar]
  • 39.Kulisevsky J., Pagonabarrage J. Cognitive Impairment in Parkinson’s Disease: Tools for Diagnosis and Assessment. Mov. Disord. 2009;24:1103–1110. doi: 10.1002/mds.22506. [DOI] [PubMed] [Google Scholar]

Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES