Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2019 Aug 19;19(16):3607. doi: 10.3390/s19163607

Joint Banknote Recognition and Counterfeit Detection Using Explainable Artificial Intelligence

Miseon Han 1, Jeongtae Kim 1,*
PMCID: PMC6719953  PMID: 31430971

Abstract

We investigated machine learning-based joint banknote recognition and counterfeit detection method. Unlike existing methods, since the proposed method simultaneously recognize banknote type and detect counterfeit detection, it is significantly faster than existing serial banknote recognition and counterfeit detection methods. Furthermore, we propose an explainable artificial intelligence method for visualizing regions that contributed to the recognition and detection. Using the visualization, it is possible to understand the behavior of the trained machine learning system. In experiments using the United State Dollar and the European Union Euro banknotes, the proposed method shows significant improvement in computation time from conventional serial method.

Keywords: banknote recognition, counterfeit banknote detection, explainable artificial intelligence, joint banknote recognition, counterfeit detection system

1. Introduction

Despite decreasing use of paper money and rapidly increasing electronic transactions, banknotes remain essential because they are easily carried, widely accepted, and reliable. For monetary transactions, it is critical that machines such as banknote counting and automatic teller machines accurately recognize banknote type and detect counterfeiting. For this reason, many studies have been conducted on banknote recognition and counterfeit detection [1,2,3,4,5].

Banknote recognition aims to identify banknote denomination and direction. Most banknote recognition methods follow the following steps [1]. First, banknote images are acquired using some sensors such as visible light and infrared sensors. Then, preprocessing crops proper image sections [1] and feature extraction is carried out to derive useful information from the cropped banknote images [1]. Finally, classification using the extracted features is performed to recognize banknote direction and classify banknotes into distinct denomination categories [1].

Conventional banknote recognition methods rely on digital image processing techniques, e.g., wavelet transforms, to extract features for classification [1]. However, designing high accuracy recognition methods is complicated because there can be many different kinds of damaged banknotes which should be recognized correctly [6]. In addition, recognition should be performed real time for most banknote recognition machines.

In contrast to digital image processing methods, machine learning methods, commonly employing convolutional neural networks (CNNs), can extract useful features without handcrafted feature extraction methods [7]. CNN based methods such as GoogleNet [8], DenseNet [9] and MobileNets [10] showed excellent performance in image classification task. Accordingly, various CNN based methods have been applied for banknote recognition and counterfeit detection [1,2,3,4,11,12,13]. Although these previous investigations were reported to perform well, computational requirements were demanding for real time implementation on embedded computers since network size were large [14].

Conventional counterfeit detection methods usually perform the detection after banknote recognition, using banknote information to extract useful features for counterfeit detection. For applying machine learning based methods, serial processing banknote recognition and counterfeit detection may require longer computational time than parallel processing [14]. Note that for most automatic teller machines, it is important to perform banknote recognition and counterfeit detection in real time without requiring large computational resource. In addition, serial banknote recognition and counterfeit detection may be error prone since recognition errors may lead to counterfeit detection failure.

To overcome the problems, we propose a novel joint banknote system to simultaneously perform banknote recognition and counterfeit detection. Because the proposed method shares convolutional layers for banknote recognition and counterfeit detection, we believe that the proposed method can be much faster than serial systems.

One of the most well known problems for machine learning classification methods is the difficulty to understand why the system generated such classification results [15]. If the machine learning system is not understandable, it may generate unexpected classification for testing data which were unseen during training. Understanding is very important to ensure confidence in the trained machine learning system. Although there have been many investigations on visualization such as classification activation map (CAM) and gradient weighted class activation map (Grad-CAM), none have been applied for machine learning based banknote recognition and counterfeit detection [16,17,18,19]. This paper is the first attempt to apply visualization methods for banknote recognition and counterfeit detection. Furthermore, we improve the traditional Grad-CAM method and propose pixel-wise Grad-CAM (pGrad-CAM) for clearer explanation for the resulting the proposed joint banknote recognition and counterfeit detection outcomes.

The remainder of this paper is organized as follows. Section 2 explains and analyzes related works in serial banknote recognition and counterfeit detection system and explainable artificial intelligence method. Section 3 summarizes contributions of this paper. Section 4 discusses two methods in detail: the proposed joint banknote recognition and counterfeit detection system and a new explainable artificial intelligence method. Section 5 experimentally demonstrates the usefulness of the proposed methods using the United State Dollar (USD) and the European Union Euro (EUR) banknotes. Section 6 summarizes and concludes the paper.

2. Related Works

2.1. Sequential Banknote Recognition and Counterfeit Detection System

One previous investigation studied a CNN based multinational banknote recognition considering size information for each banknote [5]. To reduce complexity of banknote recognition system, the method pre-classified banknote type by size, and then adopted separate CNN classifiers according to the size of the banknote [5]. Another machine learning method studied to recognize multinational banknote types using a CNN based model [3]. To handle the huge number of banknote classes, the method used pre-trained deep CNN models, AlexNet [7], GoogleNet [8], ResNet-18 and ResNet-50 [20]. Note that the pre-trained models were designed for image classification on the ImageNet database, which required an unnecessarily complicated structure for banknote recognition. Another study was conducted on banknote recognition using banknote images taken by a cellphone camera [21]. The method is based on the large frame Single Shot MultiBox Detector model [21] for banknote recognition and CNN for noise reduction [12]. Although the method showed higher performance than MobilNets [10] and faster R-CNN for banknote recognition, it suffered from real-time computation [12].

For counterfeit banknote detection, in addition to a visible light image, additional modality images such as infrared, ultraviolet, and magnetic images are often used because they provide useful information to help detect counterfeit notes [1]. Under the assumption that banknote type was correctly determined, features useful to detect counterfeit banknotes are extracted from manually predetermined regions of interest. Several methods have been studied for feature extraction, including bit plane slicing and canny edge detection [22] and luminance histogram and gray level co-occurrence matrix (GLCM) [1,23]. Using the features, counterfeit detection is performed by methods such as template matching [24] and key point matching [1,25]. Since handcrafted counterfeit detection methods require extensive efforts whenever new counterfeit types appear, machine learning based methods that can automatically learn new counterfeit types are highly desirable [4]. One previous investigation studied a CNN based counterfeit detection and proved performance using counterfeits made by general-purpose scanners [4]. Although the method achieved 100% detection accuracy [4], it included very few the Republic of Korea 10,000 Won banknotes and only considered counterfeit banknotes copied by general-purpose scanners. Another investigation studied detection of counterfeit Indian banknote using VGG16 based CNN [13]. Although the detection was successful, dataset was not sufficient to verify the performance of the method [13].

We summarize the pros and cons of conventional machine learning based sequential banknote recognition and counterfeit detection methods in comparison with joint method which we propose in this investigation. Conventional serial methods have two sequential neural networks for banknote recognition and counterfeit detection, respectively. It is relatively easy to train the networks because the loss function for each network has only single term. However, the sequential method usually requires longer computation time than the joint method due to its sequential nature [14]. The joint method which we propose has the advantage of fast computation thanks to simultaneous banknote recognition and counterfeit detection. Moreover, we design a compact CNN to accomplish faster joint banknote recognition and counterfeit detection. The fast computation of the joint method is achieved at the cost of more elaborated learning since training joint networks requires tuning of weight parameters between loss functions. One may come up with a joint method using well known CNN models for image classification task such as GoogleNet [8], DenseNet [9], and MobileNets [10]. To do that, one can modify the softmax layer of the models for joint banknote recognition and counterfeit detection and redesign loss function for training. Such modified models are very slow since the size of network is huge. Note that the CNN models are designed not for banknote recognition but for image classification task [8,9,10]. Table 1 summarizes pros and cons of sequential method, joint banknote recognition and counterfeit detection using models for image classification and proposed joint banknote recognition and counterfeit detection method.

Table 1.

A summary comparison for pros and cons.

Methods Pros Cons
Sequential method - Relatively easy to train - Relatively long inference time
Joint method using CNN for image classification - - Extremely slow
- Relatively difficult to train
Proposed method - Fast inference time - Relatively difficult to train

2.2. Grad-CAM

Deep learning systems are generally difficult to interpret why the system generated such prediction [15]. It is desirable that the system is interpretable because inexplainable systems could generate unexpected predictions, and certainty is critical for banknote recognition and counterfeit detection. One of the most useful methods to interpret image based deep learning system is a visualization technique that highlights input image regions that contributed to the prediction. Among many visualization methods, Grad-CAM [16] is one of the most widely accepted methods. The Grad-CAM generates a localization map that highlights regions with large influence on the logit value from the last convolutional layer of CNN because feature maps derived from the layer have the most distinguishable features while retaining the spatial information.

The Grad-CAM method first computes the importance weight αkc as follows:

αkc=1MNijycAijk, (1)

where yc is the logit value for target class c and Ak represents M×N size k-th feature map of the last convolutional layer. The importance weight αkc is computed by average pooling over the feature map Ak. Using the importance weights, Grad-CAM LGradCAMc as follows:

LGrad-CAMc=ReLU(kαkcAk). (2)

Note that the Grad-CAM is generated by computing weighted sum of feature maps where weight for each feature map is computed by the average of gradient values on the feature map. Grad-CAM considers only positive influences on the class of interest by applying ReLU operation after combining feature maps using the importance weights.

Although Grad-CAM has been successfully applied for various applications [26,27,28], we think that the method should be improved because Grad-CAM generates empty activation maps for some cases. This happens when almost αkc values are negative values, because negative gradients values cancel positive gradient values during average-pooling. Figure 1 shows such example images which have predicted probabilities higher than 0.99 for both banknote recognition and counterfeit detection. Although the images can be clearly classified, as the high probabilities suggest, Grad-CAM activation maps were empty.

Figure 1.

Figure 1

Grad-CAM results for some example banknotes. The most left column to the third column show visible, infrared transmission, and infrared reflection images of banknotes. The forth column and the most right column show Grad-CAM results for banknote recognition and counterfeit detection, reflectively.

3. Contributions

The novelties of the proposed method are joint banknote recognition and counterfeit detection system and explainable artificial intelligence method for banknote recognition and counterfeit detection. To our knowledge, the proposed method is the first method which simultaneously classified banknote denomination, and counterfeit detection. Previous methods rely on denomination results to detect counterfeit banknote, which is slow. For the joint banknote recognition, and counterfeit detection, we use aligned visible, infrared reflection, and infrared transmission images as input data. We also proposed a novel loss function which combines three loss functions for banknote denomination, banknote direction, and counterfeit detection. Furthermore, our research is the first attempt to ensure the performance of banknote recognition method using explainable artificial intelligence. To do that, we propose a novel visualization method for banknote recognition and counterfeit detection system that overcomes shortcomings of an existing method. Note that we do not propose a novel CNN structure for banknote recognition system. The novelties of the proposed method are joint banknote recognition and counterfeit detection system and a visualization method to explain the system.

4. Methods

4.1. Joint Banknote Recognition and Counterfeit Detection System

Conventional systems sequentially perform banknote recognition and counterfeit detection since counterfeit detection relies on the result of recognition [1]. Banknote recognition requires not only classifying the banknote denomination, but also its direction, because directional information is used for tasks such as banknote serial number recognition [1]. Figure 2 shows direction can be classified into four categories A, B, C, and D: front view, upside down (flipped), back to front (flopped), and flipped and flopped, respectively.

Figure 2.

Figure 2

Possible banknote directions.

Counterfeit detection detects forged banknotes based on features that only occur on genuine banknotes. Figure 3 shows examples of genuine features. Genuine EUR banknotes include an infrared transmission image with a center strip, which is not visible on the counterfeit banknote image while genuine USD banknotes have a center dashed strip on the infrared reflection image. To detect counterfeit features, some previous methods use digital image processing techniques. It is also possible to detect using machine learning based method. However, all previous methods use the information of banknote type to detect features for counterfeit detection.

Figure 3.

Figure 3

Different modality images. Leftmost column shows visible, center column shows infrared transmission, and rightmost column shows infrared reflection images.

As pointed in the above, sequential banknote recognition and counterfeit detection can be slow since shallow networks are usually faster than deep networks [14]. To overcome the problem, we propose a novel joint banknote system that can simultaneously perform banknote recognition and counterfeit detection, as shown in Figure 4. The joint system consists of two convolutional layers, two max-pooling layers, and two fully connected layers followed by a softmax layer. We also used a batch normalization layer to avoid internal covariance shift [29], and the well known rectified linear unit (ReLU) nonlinear activation layer. Although there exist sophisticated nonlinear activation functions such as leaky ReLU [30], ReLU was sufficient to accomplish desired performance for the proposed network. Furthermore, ReLU is known to be more robust to noise than leaky ReLU [31,32].

Figure 4.

Figure 4

Joint banknote recognition and counterfeit detection system.

We use the first convolutional layer with 32 feature maps and 3×3 convolution kernels to extract the features from three channel input data with a visible, an infrared reflection, and an infrared transmission images. After a batch normalization and a ReLU layer, convolved feature maps are subsampled by max-pooling with stride 2. Then, we use the subsampled feature maps as the input of the second convolutional layer with 32 feature maps and 3×3 convolution kernels. The output of the second convolutional layer is connected to first fully-connected layer with 32 nodes after a series of batch normalization, ReLU, and max-pooling layer with stride 2. The second fully connected layer generates logit values for banknote denomination, direction, and counterfeit detection. Then, the logit values are converted into probabilities using softmax layers as follows:

pic=euiieui,i=0,1, (3)

where pic denotes the probability of the prediction for counterfeit detection and

pin=eviievi,i=0,,N1, (4)

where pin denotes the probability of the prediction for banknote denomination and N is the number of banknote denominations for a country.

pid=ewiiewi,i=0,,3, (5)

where pid is the probability of the prediction for banknote direction.

The proposed system was trained by minimizing a loss function combining Ldet for counterfeit banknote detection, Lden for banknote denomination, and Ldir for banknote direction. Mathematically, the total loss function Ltot is defined as follows:

Ltot=λ1Ldet+λ2Lden+λ3Ldir, (6)

where λ1, λ2, and λ3 control weights for the three loss functions. In addition, Ldet is defined as follows:

Ldet=i=01qiclogpic, (7)

where qic is the true labeled probability of input banknote data, and pic is the corresponding prediction from the network. The denomination loss function is defined by

Lden=i=0N1qinlogpin, (8)

where qin represents labeled probability of each denomination, pin denotes the corresponding prediction and N is the number of banknote denominations for a country. Finally, the direction loss function is defined by

Ldir=i=03qidlogpid, (9)

where i is one of A, B, C, and D directions, qid is labeled probability for each direction and pid is the corresponding predicted probability.

The joint banknote recognition and counterfeit detection system predicts if the banknote is genuine or counterfeit. The joint system classifies banknote denomination where the number of classes depends on the number of different banknotes. The proposed system also predicts banknote direction. Note that system architecture can vary slightly for different national banknotes.

4.2. Explainable Artificial Intelligence

To resolve the empty activation map problem with Grad-CAM mentioned in Section 2.2, we propose a pixel-wise gradient weighted class activation map (pGrad-CAM) for visualization. The idea of pGrad-CAM is that if a feature map has both positive and negative gradients on a testing logit value, regions of positive gradients should be taken into consideration pixel-wisely even if the average of the gradient values is negative. Thus, we first compute weight map Wkc of the k-th feature for c-th class as follows:

Wkc=ReLU(GσAkyc), (10)

where yc is a logit for a target class c, Ak means k-th feature map, * means convolution operator and Gaussian smoothing operator Gσ is defined as follows:

Gσ(i,j)=12πσ2ei2+j22σ2 (11)

where σ is a standard deviation of the distribution and i and j are pixel locations over the feature map. Using Gaussian smoothing operator, we would like to reduce the difference between intensities of adjacent pixels and extract more global features.

Using a linearly weighted combination between influences and feature maps, we compute pGrad-CAM as follows:

LpGrad-CAMc=k(WkcAk), (12)

where ⊙ is the pixel-wise product operation. Since Wkc never becomes all zero values, if positive gradient values exists, then pGrad-CAM never generate an empty map.

Figure 5 shows the block diagram to obtain pGrad-CAM. We first calculate the influence of the feature map extracted by the last convolutional layer on a logit through the gradient of the logit with respect to the feature map. We can identify the influence of each pixel on the logit yc due to pixel-wise operation. In addition, pGrad-CAM can highlight valid features using positive gradients without counterbalancing effects from negative influences on the logit. Using pGrad-CAM, we can understand how the network determines predictions and decides whether the model has high or low confidence.

Figure 5.

Figure 5

pGrad-Cam flow.

5. Experimental Results

To evaluate the performance of the proposed method in comparison with sequential banknote recognition and counterfeit detection method, we conducted an experiment with EUR and USD banknotes. Genuine banknote images and counterfeit EUR and USD banknote images were acquired by Puloon Technology (Republic of Korea) using EagleEye10. The counterfeit EUR and USD images were acquired from the European Central Bank and the Federal Reserve Bank, respectively. We used 60×120 visible, infrared transmission, and an infrared reflection images for a total of 23,955 ×3 and 45,055 ×3 EUR and USD banknotes, respectively.

We implemented the proposed method as well as a sequential banknote recognition and counterfeit detection method. Further, for comparison purpose, we implemented a joint banknote recognition and counterfeit detection method using well known CNN for image classification. We selected GoogleNet which is composed of 22 layers [8] as the comparison method because the network showed high performance and fast inference time for image classification task [33]. We modified the final softmax layer of GoogleNet for the joint banknote recognition and counterfeit detection to compute the combined loss function defined in (6). We call this method by joint GoogleNet method.

We divided all banknote images into training, validation, and test datasets at 9:0.5:0.5 ratio, respectively. For training data, we flipped the counterfeit banknote images horizontally, vertically, and both horizontally and vertically for data augmentation. As the result, the augmented counterfeit dataset is four times larger than original counterfeit dataset. Table 2 summarizes dataset sizes for each banknote type. The training dataset of the EUR banknotes comprised 18,321 genuine and 3476 counterfeit banknotes (including augmentation) while the training dataset of the USD banknotes comprised 33,583 genuine and 7520 counterfeit banknotes (including augmentation).

Table 2.

Banknote datasets.

Nation Denomination The Number of Dataset
Train Validation Test
Series Type Genuine Counterfeit Genuine Counterfeit Genuine Counterfeit
EUR First 5 EUR 3266 8 182 1 182 1
10 EUR 744 76 42 2 42 2
20 EUR 634 140 36 3 36 3
50 EUR 864 536 49 8 49 8
100 EUR 3110 216 173 3 173 3
200 EUR 824 1364 46 20 46 20
500 EUR 967 644 54 9 54 9
Second 5 EUR 1535 112 86 2 86 2
10 EUR 2304 40 129 1 129 1
20 EUR 1371 268 77 4 77 4
50 EUR 2702 72 151 1 151 1
Total 18,321 3476 1025 54 1025 54
USD 1 USD 1735 0 97 0 97 0
2 USD 3780 0 210 0 210 0
5 USD 2574 0 143 0 143 0
10 USD 5298 1664 295 24 295 24
20 USD 8369 5436 466 76 466 76
50 USD 8263 340 460 5 460 5
100 USD 3564 80 198 2 198 2
Total 33,583 7520 1869 107 1869 107

We implemented the proposed and the sequential methods using Tensorflow and tested the trained model on an embedded NVIDIA Jet AGX system (NVIDIA, USA). For training network, we determined the weights of the loss function manually as λ1=2, λ2=1, and λ3=1. To avoid boundary artifacts, we use convolution with valid option instead of convolution with same or full option [34]. In addition, to avoid overfitting, we applied 50% dropout between the last convolutional layer and the first fully connected layer. We applied dropout only between the last convolutional layer and the first fully connected layer because the second fully connected layer does not have many connections. Although one may apply dropout in convolutional layer too, it has been known that performance enhancing of dropout in convolution layer is minimal [35]. Note that the fully connected dropout is known as the best performer on a small architecture [36]. We trained the joint banknote recognition and counterfeit detection system using ADAM optimizer with 0.0001 adaptive learning rate, 0.9 decay factor, and 512 batch size for EUR while using ADAM optimizer with 0.0001 fixed learning rate and 512 batch size for USD [37]. We terminated training when training and validation losses were below the pre-determined threshold (0.005). Figure 6 shows the changes of average batch losses for training and validation datasets during training. The final average training and validation losses were 6.97×105 and 1.87×104 for EUR banknote network while 4.18×105 and 7.89×104 for USD banknote network.

Figure 6.

Figure 6

Average batch losses of convergence graphs.

Table 3 shows banknote recognition and counterfeit detection accuracies which are defined as the ratio of the number of correctly recognized and correctly classified into genuine or counterfeit to the total number of tested banknote images. For EUR and USD banknotes, all methods achieved 100% accuracies, which imply machine learning based method can perform very well for banknote recognition and counterfeit detection. This result is not surprising since machines which deal with currency must show extremely high accuracy. Note that we have tested real banknote images for this experiments. Needless to say, if the methods were tested using damaged banknotes, the proposed method as well as other methods may not show 100% accuracy. Investigation of performance for damaged banknote images is deferred to future study.

Table 3.

Accuracy of banknote system.

Nation Dataset Number of Banknotes (Counterfeit) Number of Well-Classified (Counterfeit) Accuracy (%)
Sequential Method Joint GoogleNet Proposed Method
EUR Train and validation 22,876 (3530) 22,876 (3530) 22,876/22,876 (100) 22,876/22,876 (100) 22,876/22,876 (100)
Test 1079 (54) 1079 (54) 1079/1079 (100) 1079/1079 (100) 1079/1079 (100)
USD Train and validation 43,079 (7627) 43,079 (7627) 43,079/43,079 (100) 43,079/43,079 (100) 43,079/43,079 (100)
Test 1976 (107) 1976 (107) 1976/1976 (100) 1976/1976 (100) 1976/1976 (100)

Although all methods showed the same accuracy, the proposed method was markedly faster than the other methods. Table 4 shows mean and variance of the computational time for 1000 times executions for each method. We used the same network as in the proposed method twice for the sequential banknote recognition and counterfeit detection method. The preprocessing time is computation time for image resizing and normalization. The total computational time of the sequential method was about 11.69 ms. Except for preprocessing, the sequential method required approximately 4.18 ms and 3.94 ms for banknote recognition and counterfeit detection, respectively. The joint GoogleNet required similar preprocessing. However, computation time for joint banknote recognition and counterfeit detection was about 947.12 ms, which is impractical to be used for banknote recognition. This is due to the joint GoogleNet has lots of layers and weight parameters, which may be necessary for image classification but not for banknote recognition. On the contrary, the total computational time of the proposed method is only about 8.09 ms. The proposed method took about 4.36 ms for the joint banknote recognition and counterfeit detection. One may think that performance improvement of 3.6 ms (difference between means of the sequential method and the proposed method) may not be important. However, the performance improvement is more than 30%, which can be critical for realtime implementation on an embedded machine.

Table 4.

Mean and variance of inference time.

Model Processing Time on Average (variance)
Preprocessing Banknote Recognition Counterfeit Detection Total
Sequential 3.57 ms (0.46) 4.18 ms (0.86) 3.94 ms (0.69) 11.69 ms (3.02)
Joint GoogleNet 3.53 ms (0.12) 947.12 ms (2817.44) 950.65 ms (2804.65)
Proposed 3.73 ms (0.49) 4.36 ms (0.72) 8.09 ms (1.18)

Even though the proposed system achieved 100% accuracy for both banknote classification and counterfeit detection and improved speed compared with other methods, we believe it is essential for the system to be well explained to ensure confidence that the system will not generate unexpected outcomes for unseen data. To do that, we generated activation maps and identified regions with large effects on the prediction. Table 5 shows the activation maps of the joint system using Grad-CAM and pGrad-CAM.

Table 5.

Grad-CAM and pGrad-CAM results.

Banknote Input Images Banknote Recognition Counterfeit Detection
Visible Infrared Transmission Infrared Reflection Grad-CAM pGrad-CAM Grad-CAM pGrad-CAM
20 EUR first series graphic file with name sensors-19-03607-i005a.jpg graphic file with name sensors-19-03607-i005b.jpg graphic file with name sensors-19-03607-i005c.jpg graphic file with name sensors-19-03607-i005d.jpg graphic file with name sensors-19-03607-i005e.jpg graphic file with name sensors-19-03607-i005f.jpg graphic file with name sensors-19-03607-i005g.jpg
20 EUR second series graphic file with name sensors-19-03607-i005h.jpg graphic file with name sensors-19-03607-i005i.jpg graphic file with name sensors-19-03607-i005j.jpg graphic file with name sensors-19-03607-i005k.jpg graphic file with name sensors-19-03607-i005l.jpg graphic file with name sensors-19-03607-i005m.jpg graphic file with name sensors-19-03607-i005n.jpg
200 EUR graphic file with name sensors-19-03607-i005o.jpg graphic file with name sensors-19-03607-i005p.jpg graphic file with name sensors-19-03607-i005q.jpg graphic file with name sensors-19-03607-i005r.jpg graphic file with name sensors-19-03607-i005s.jpg graphic file with name sensors-19-03607-i005t.jpg graphic file with name sensors-19-03607-i005u.jpg
500 EUR graphic file with name sensors-19-03607-i005v.jpg graphic file with name sensors-19-03607-i005w.jpg graphic file with name sensors-19-03607-i005x.jpg graphic file with name sensors-19-03607-i005y.jpg graphic file with name sensors-19-03607-i005z.jpg graphic file with name sensors-19-03607-i005{.jpg graphic file with name sensors-19-03607-i005--.jpg
1 USD graphic file with name sensors-19-03607-i005}.jpg graphic file with name sensors-19-03607-i005~.jpg graphic file with name sensors-19-03607-i005aa.jpg graphic file with name sensors-19-03607-i005ba.jpg graphic file with name sensors-19-03607-i005ca.jpg graphic file with name sensors-19-03607-i005da.jpg graphic file with name sensors-19-03607-i005ea.jpg
2 USD graphic file with name sensors-19-03607-i005fa.jpg graphic file with name sensors-19-03607-i005ga.jpg graphic file with name sensors-19-03607-i005ha.jpg graphic file with name sensors-19-03607-i005ia.jpg graphic file with name sensors-19-03607-i005ja.jpg graphic file with name sensors-19-03607-i005ka.jpg graphic file with name sensors-19-03607-i005la.jpg
50 USD graphic file with name sensors-19-03607-i005ma.jpg graphic file with name sensors-19-03607-i005na.jpg graphic file with name sensors-19-03607-i005oa.jpg graphic file with name sensors-19-03607-i005pa.jpg graphic file with name sensors-19-03607-i005qa.jpg graphic file with name sensors-19-03607-i005ra.jpg graphic file with name sensors-19-03607-i005sa.jpg
100 USD graphic file with name sensors-19-03607-i005ta.jpg graphic file with name sensors-19-03607-i005ua.jpg graphic file with name sensors-19-03607-i005va.jpg graphic file with name sensors-19-03607-i005wa.jpg graphic file with name sensors-19-03607-i005xa.jpg graphic file with name sensors-19-03607-i005ya.jpg graphic file with name sensors-19-03607-i005.jpg

In Table 5, from the most left column to the third column show visible, infrared transmission and infrared reflection images of banknotes. The fourth and fifth columns show Grad-CAM and pGrad-CAM results for banknote recognition while the sixth column and the most right column show Grad-CAM and pGrad-CAM results for counterfeit detection. In the Grad-CAM and the pGrad-CAM results, more reddish regions correspond to higher influence on the class whereas more bluish regions correspond to lower influence on the class.

Although Grad-CAM generated explainable results for most cases, the method failed for some EUR and USD banknotes as shown in Table 5. The probabilities of all banknotes in Table 5 have higher than 0.99 for both banknote recognition and counterfeit detection. However, Grad-CAM did not show anything for some cases (i.e., all zero values) such as 20 EURs, 200 EUR, 1 USD, and 2 USD for banknote recognition and 50 USD and 100 USD for counterfeit detection. In 500 EUR banknotes, although Grad-CAM results for banknote recognition have some colored pixels, the effects seem to be too restricted to understand the behavior of the network. Similar phenomena occur for 1 USD and 2 USD banknotes for counterfeit detection.

Unlike Grad-CAM, pGrad-CAM for 20 EUR first series and 20 EUR second series highlights the upper digits as very important clues to classify banknote denomination, which we believe reasonable. For 200 EUR banknote, pGrad-CAM highlights regions on the upper digits and the center of door in a visible image as important features for banknote recognition. For 500 EUR banknote recognition, pGrad-CAM highlights regions that contain important features such as regions near the left rectangular box. For 1 USD and 2 USD banknotes, pGrad-CAM highlights the center portrait and features near sides of the banknote. For counterfeit detection, Grad-CAM highlights similar regions highlighted in pGrad-CAM for EUR banknotes. However, Grad-CAM appears to emphasize irrelevant regions for USD banknotes such as near the margin.

Table 6 shows averaged Grad-CAM and pGrad-CAM results for banknote denomination, direction, and counterfeit detection. For EUR banknote denomination, it seems that averaged Grad-CAM highlights less important regions such as the upper right rectangular box for 20 EUR first series banknote and the bottom right rectangular box for 100 EUR banknote than pGrad-CAM. The pGrad-CAM stresses important features such as the upper digits for 20 EUR first series banknote and the door in the lower left region in 100 EUR banknote. For USD banknote denomination and EUR counterfeit detection, both Grad-CAM and pGrad-CAM highlight similar regions. For 10 USD banknote, Grad-CAM did not indicate lower left regions that have counterfeit features. In contrast, pGrad-CAM seems to highlight regions that contain important features for banknote recognition and counterfeit detection. We think this is because pGrad-CAM is based on the pixelwise effects of feature maps on prediction, which may provide more accurate information about regions that effect prediction.

Table 6.

Average Grad-CAM and pGrad-CAM results.

Banknote Method Input Example Images Explainable Artificial Intelligence
Visible Infrared Transmission Infrared Reflection Banknote Denomination Banknote Direction Counterfeit Detection
20 EUR first series Grad-CAM graphic file with name sensors-19-03607-i006a.jpg graphic file with name sensors-19-03607-i006b.jpg graphic file with name sensors-19-03607-i006c.jpg graphic file with name sensors-19-03607-i006d.jpg graphic file with name sensors-19-03607-i006e.jpg graphic file with name sensors-19-03607-i006f.jpg
pGrad-CAM graphic file with name sensors-19-03607-i006g.jpg graphic file with name sensors-19-03607-i006h.jpg graphic file with name sensors-19-03607-i006i.jpg graphic file with name sensors-19-03607-i006j.jpg graphic file with name sensors-19-03607-i006k.jpg graphic file with name sensors-19-03607-i006l.jpg
100 EUR Grad-CAM graphic file with name sensors-19-03607-i006m.jpg graphic file with name sensors-19-03607-i006n.jpg graphic file with name sensors-19-03607-i006o.jpg graphic file with name sensors-19-03607-i006p.jpg graphic file with name sensors-19-03607-i006q.jpg graphic file with name sensors-19-03607-i006r.jpg
pGrad-CAM graphic file with name sensors-19-03607-i006s.jpg graphic file with name sensors-19-03607-i006t.jpg graphic file with name sensors-19-03607-i006u.jpg graphic file with name sensors-19-03607-i006v.jpg graphic file with name sensors-19-03607-i006w.jpg graphic file with name sensors-19-03607-i006x.jpg
10 USD Grad-CAM graphic file with name sensors-19-03607-i006y.jpg graphic file with name sensors-19-03607-i006z.jpg graphic file with name sensors-19-03607-i006{.jpg graphic file with name sensors-19-03607-i006--.jpg graphic file with name sensors-19-03607-i006}.jpg graphic file with name sensors-19-03607-i006~.jpg
pGrad-CAM graphic file with name sensors-19-03607-i006aa.jpg graphic file with name sensors-19-03607-i006ba.jpg graphic file with name sensors-19-03607-i006ca.jpg graphic file with name sensors-19-03607-i006da.jpg graphic file with name sensors-19-03607-i006ea.jpg graphic file with name sensors-19-03607-i006fa.jpg
20 USD Grad-CAM graphic file with name sensors-19-03607-i006ga.jpg graphic file with name sensors-19-03607-i006ha.jpg graphic file with name sensors-19-03607-i006ia.jpg graphic file with name sensors-19-03607-i006ja.jpg graphic file with name sensors-19-03607-i006ka.jpg graphic file with name sensors-19-03607-i006la.jpg
pGrad-CAM graphic file with name sensors-19-03607-i006ma.jpg graphic file with name sensors-19-03607-i006na.jpg graphic file with name sensors-19-03607-i006oa.jpg graphic file with name sensors-19-03607-i006pa.jpg graphic file with name sensors-19-03607-i006qa.jpg graphic file with name sensors-19-03607-i006.jpg

6. Conclusions

This paper proposed a simultaneous banknote recognition and counterfeit detection system, with a related explainable artificial intelligence method. Experiments using USD and EUR banknotes confirmed the proposed method achieved significantly faster computation than the conventional sequential method while retaining 100% banknote recognition and counterfeit detection accuracy. The proposed pGrad-CAM structure also explained proposed network system behavior better than conventional Grad-CAM. Therefore, we believe that the proposed method will be useful for practical banknote recognition and counterfeit detection.

Acknowledgments

The authors are very grateful to Puloon Technology in the Republic of Korea for providing us experimental data set.

Author Contributions

M.H. conducted experimental studies and wrote the draft. J.K. conceptualized the proposed method, supervised the simulation and experimental studies and improved the draft.

Funding

This research was supported by the Technology development Program (S2467392) funded by the Ministry of SMEs and Startups (MSS, Korea) and by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT & Future Planning (2017R1A2B4004231).

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Woo Lee J., Hong H., Wan Kim K., Ryoung Park K. A Survey on Banknote Recognition Methods by Various Sensors. Sensors. 2017;17:313. doi: 10.3390/s17020313. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Park Y., Kwon S., Pham T., Park K., Jeong D., Yoon S. A high performance banknote recognition system based on a one-dimensional visible light line sensor. Sensors. 2015;15:14093–14115. doi: 10.3390/s150614093. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Pham T.D., Nguyen D.T., Park C., Park K.R. Deep Learning-Based Multinational Banknote Type and Fitness Classification with the Combined Images by Visible-Light Reflection and Infrared-Light Transmission Image Sensors. Sensors. 2019;19:792. doi: 10.3390/s19040792. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Lee S.H., Lee H.Y. Counterfeit Bill Detection Algorithm using Deep Learning. Int. J. Appl. Eng. Res. 2018;13:304–310. [Google Scholar]
  • 5.Pham T., Lee D., Park K. Multi-national banknote classification based on visible-light line sensor and convolutional neural network. Sensors. 2017;17:1595. doi: 10.3390/s17071595. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Sarfraz M. An intelligent paper currency recognition system. Procedia Comput. Sci. 2015;65:538–545. doi: 10.1016/j.procs.2015.09.128. [DOI] [Google Scholar]
  • 7.Krizhevsky A., Sutskever I., Hinton G.E. Advances in Neural Information Processing Systems. The MIT Press; Cambridge, MA, USA: 2012. Imagenet classification with deep convolutional neural networks; pp. 1097–1105. [Google Scholar]
  • 8.Szegedy C., Liu W., Jia Y., Sermanet P., Reed S., Anguelov D., Erhan D., Vanhoucke V., Rabinovich A. Going deeper with convolutions; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Boston, MA, USA. 7–12 June 2015; pp. 1–9. [Google Scholar]
  • 9.Iandola F., Moskewicz M., Karayev S., Girshick R., Darrell T., Keutzer K. Densenet: Implementing efficient convnet descriptor pyramids. arXiv. 20141404.1869 [Google Scholar]
  • 10.Howard A.G., Zhu M., Chen B., Kalenichenko D., Wang W., Weyand T., Andreetto M., Adam H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv. 20171704.04861 [Google Scholar]
  • 11.Ren Y. Ph.D. Thesis. Auckland University of Technology; Auckland, New Zealand: 2017. Banknotes Recognition in Real Time Using ANN. [Google Scholar]
  • 12.Zhang Q. Ph.D. Thesis. Auckland University of Technology; Auckland, New Zealand: 2018. Currency Recognition Using Deep Learning. [Google Scholar]
  • 13.Navya Krishna G., Sai Pooja G., Naga Sri Ram B., Yamini Radha V., Rajarajeswari P. Recognition of fake currency note using convolutional neural networks. Int. J. Innov. Technol. Explor. Eng. 2019;8:58–63. [Google Scholar]
  • 14.Ba J., Caruana R. Advances in Neural Information Processing Systems. The MIT Press; Cambridge, MA, USA: 2014. Do deep nets really need to be deep? pp. 2654–2662. [Google Scholar]
  • 15.Samek W., Wiegand T., Müller K.R. Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. arXiv. 20171708.08296 [Google Scholar]
  • 16.Selvaraju R.R., Cogswell M., Das A., Vedantam R., Parikh D., Batra D. Grad-CAM: Visual explanations from deep networks via gradient-based localization; Proceedings of the IEEE International Conference on Computer Vision; Venice, Italy. 22–29 October 2017; pp. 618–626. [Google Scholar]
  • 17.Andreotti F., Phan H., De Vos M. Visualising convolutional neural network decisions in automatic sleep scoring; Proceedings of the Joint Workshop on Artificial Intelligence in Health (AIH) 2018; Stockholm, Sweden. 13–14 July 2018; pp. 70–81. [Google Scholar]
  • 18.Woo S., Park J., Lee J.Y., So Kweon I. CBAM: Convolutional block attention module; Proceedings of the European Conference on Computer Vision (ECCV); Munich, Germany. 8–14 September 2018; pp. 3–19. [Google Scholar]
  • 19.Graziani M., Andrearczyk V., Müller H. Visual Interpretability for Patch-Based Classification of Breast Cancer Histopathology Images. [(accessed on 19 August 2019)];2018 Available online: https://openreview.net/forum?id=S1PTal9sz.
  • 20.He K., Zhang X., Ren S., Sun J. Deep residual learning for image recognition; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA. 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  • 21.Liu W., Anguelov D., Erhan D., Szegedy C., Reed S., Fu C.Y., Berg A.C. European Conference on Computer Vision. Springer; Amsterdam, The Netherlands: 2016. SSD: Single shot multibox detector; pp. 21–37. [Google Scholar]
  • 22.Alshayeji M.H., Al-Rousan M., Hassoun D.T. Detection method for counterfeit currency based on bit-plane slicing technique. Int. J. Multimed. Ubiquitous Eng. 2015;10:225–242. doi: 10.14257/ijmue.2015.10.11.22. [DOI] [Google Scholar]
  • 23.Bhavani R., Karthikeyan A. A novel method for counterfeit banknote detection. Int. J. Comput. Sci. Eng. 2014;2:165–167. [Google Scholar]
  • 24.Ambadiyil S., Reddy T., Teja B., Pillai V. Banknote authentication using normalized cross correlation method. Discovery. 2015;44:166–172. [Google Scholar]
  • 25.Lamsal S., Shakya A. Counterfeit paper banknote identification based on color and texture; Proceedings of the IOE Graduate Conference; Lalitpur, Nepal. 11–12 December 2015; pp. 160–168. [Google Scholar]
  • 26.Ching T., Himmelstein D.S., Beaulieu-Jones B.K., Kalinin A.A., Do B.T., Way G.P., Ferrero E., Agapow P.M., Zietz M., Hoffman M.M., et al. Opportunities and obstacles for deep learning in biology and medicine. J. R. Soc. Interface. 2018;15 doi: 10.1098/rsif.2017.0387. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Han S.S., Kim M.S., Lim W., Park G.H., Park I., Chang S.E. Classification of the clinical images for benign and malignant cutaneous tumors using a deep learning algorithm. J. Investig. Dermatol. 2018;138:1529–1538. doi: 10.1016/j.jid.2018.01.028. [DOI] [PubMed] [Google Scholar]
  • 28.Martinel N., Foresti G.L., Micheloni C. Wide-slice residual networks for food recognition; Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV); Lake Tahoe, NV, USA. 12–15 March 2018; pp. 567–576. [Google Scholar]
  • 29.Ioffe S., Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv. 20151502.03167 [Google Scholar]
  • 30.Maas A.L., Hannun A.Y., Ng A.Y. Rectifier nonlinearities improve neural network acoustic models; Proceedings of the 2013 International Conference on Machine Learning (ICML); Atlanta GA, USA. 16–21 June 2013; p. 3. [Google Scholar]
  • 31.Sivadas S., Wu Z., Bin M. Investigation of parametric rectified linear units for noise robust speech recognition; Proceedings of the Sixteenth Annual Conference of the International Speech Communication Association; Dresden, Germany. 6–10 September 2015. [Google Scholar]
  • 32.Clevert D.A., Unterthiner T., Hochreiter S. Fast and accurate deep network learning by exponential linear units (elus) arXiv. 20151511.07289 [Google Scholar]
  • 33.Bianco S., Cadene R., Celona L., Napoletano P. Benchmark analysis of representative deep neural network architectures. IEEE Access. 2018;6:64270–64277. doi: 10.1109/ACCESS.2018.2877890. [DOI] [Google Scholar]
  • 34.Goodfellow I., Bengio Y., Courville A. Deep Learning. MIT Press; Cambridge, MA, USA: 2016. [(accessed on 19 August 2019)]. Available online: http://www.deeplearningbook.org. [Google Scholar]
  • 35.Srivastava N., Hinton G., Krizhevsky A., Sutskever I., Salakhutdinov R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014;15:1929–1958. [Google Scholar]
  • 36.Wu H., Gu X. Towards dropout training for convolutional neural networks. Neural Netw. 2015;71:1–10. doi: 10.1016/j.neunet.2015.07.007. [DOI] [PubMed] [Google Scholar]
  • 37.Kingma D.P., Ba J. Adam: A method for stochastic optimization. arXiv. 20141412.6980 [Google Scholar]

Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES