Skip to main content
The Scientific World Journal logoLink to The Scientific World Journal
. 2014 Jan 29;2014:168275. doi: 10.1155/2014/168275

Gait Correlation Analysis Based Human Identification

Jinyan Chen 1,*
PMCID: PMC3925574  PMID: 24592144

Abstract

Human gait identification aims to identify people by a sequence of walking images. Comparing with fingerprint or iris based identification, the most important advantage of gait identification is that it can be done at a distance. In this paper, silhouette correlation analysis based human identification approach is proposed. By background subtracting algorithm, the moving silhouette figure can be extracted from the walking images sequence. Every pixel in the silhouette has three dimensions: horizontal axis (x), vertical axis (y), and temporal axis (t). By moving every pixel in the silhouette image along these three dimensions, we can get a new silhouette. The correlation result between the original silhouette and the new one can be used as the raw feature of human gait. Discrete Fourier transform is used to extract features from this correlation result. Then, these features are normalized to minimize the affection of noise. Primary component analysis method is used to reduce the features' dimensions. Experiment based on CASIA database shows that this method has an encouraging recognition performance.

1. Introduction

Biometrics is a technology that makes use of the physiological or behavioral characteristics to authenticate or identify people [1]. The most commonly used biometrics applications are fingerprint and iris based identification.

Human gait was firstly studied in the medical field [25]. Doctors analyzed human gait to find out whether patients had health problem. Later, researchers [5] found that just like fingerprint and iris, almost everyone had his distinctive walking style. So someone believed that gait could also be used as a biological feature to identify a person. Comparing to fingerprint and iris based identification, human gait identification has the following advantages [6]: (1) it does not require the user's interaction and (2) it can be done at a distance, as long as the gait is visible.

Current human gait identification can be divided into two categories: model-based methods and motion-based methods.

Model-based approaches aim to describe human movement using a mathematical model, for example, Cunado et al. [7] used Hough translation to extract arms, legs, and torso and use articulated pendulum to match the moving body parts. Yoo et al. [8] divide the body into head, neck, waist, leg, and arm by image segmentation and then obtained the moving curve of these body parts. Lee and Grimson [9] applied 7 ellipses to model the human body and applied the ellipses' movement features to identify human. Yam et al. [10, 11] used dynamically coupled oscillator to describe and analyze the walking and running style of a person. Tafazzoli and Safabakhsh [12] constructed movements model based on anatomical proportions then Fourier transform was used to analysis human walking style. Dupuis et al. [13] created probabilistic based gait modeling to describe the human walking.

Motion-based methods consider the human gait as a sequence of image and extract features from these images. Cheng et al. [14] employed Hidden Markov Models to analyze the relationship among these images. Chen et al. [15] used parallel HMM to describe the features of human gait. Yu et al. [16] overlapped all the images to get Gait Energy Image (GEI) and GEI was used as the features to identify humans. Similar to GEI, Fan et al. [17] took Chrono-Gait Image as the gait features. Kale et al. [18] used “frieze” patterns to extract features from image sequence and used these features to identify humans. Liu et al. [19] represent one's walking style by outermost contour.

The raw features extracted by motion-based methods usually contained several cures or images, so Gabor transform [20], Random Forest algorithm [12], and Fourier transform [21, 22] were used to reduce the dimensions of these features.

1.1. Overview of Our Approach

Unlike traditional image correlation based identification [16, 23, 24], this paper aims to build a human gait identification method not only in spatial domain but also in temporal domain. A median model is used to estimate the background. By background subtraction, we get the binary human silhouette. Every silhouette is correlated with its neighbor at horizontal, vertical, and temporal axis. The size of the correlation result is used to construct the walking feature curves and every feature curve consists of several walking cycles. The feature curve is normalized to build a standardized curve. For every image sequence, we can get totally 13 standardized curves. Primary component analysis is used to extract features from these above curves. The overview of the proposed algorithm is shown in Figure 1.

Figure 1.

Figure 1

Overview of our human identification method.

1.2. Our Contribution

The main purpose and contributions of this paper can be summarized as follows.

  1. We extract 13 feature curves from the human gait images sequence. These 13 curves described the feature of human walking.

  2. We proposed a method to construct normalized human gait feature vectors. PCA is used to reduce the dimensions of gait feature vectors.

  3. We discuss the relationship between accuracy of identification and the shift length along horizontal axis and vertical axis and try to find out the best shift length.

2. Feature Extraction

2.1. Silhouettes Sequence Extraction

In this paper, the human gait is considered as a binary image. Gauss background model is used to estimate the background. We use the method proposed by Wang et al. [25] to subtract the background from the foreground of the image as follows:

D(a,b)=12(a+1)(b+1)(a+1)+(b+1)·2(256a)(256b)(256a)+(256b), (1)

D(a, b) is the difference between point a (pixel in foreground image) and point b (pixel in background image).

2.2. About the Image Sequence Correlation

Image correlation is a shift-invariant feature method and was first proposed by Otsu and Kurita [26]. Let f(R) be D-dimensions data with R = (x 1, x 2,…, x D); then the correlation function is defined as

ψ(f(R),Si)=f(R)f(R+Si)dR, (2)

where (S i,1,…, S i,D) is the shift vector.

Given a human gait image sequence, f(R t) means an image in the sequence whose frame index is t. If S i = (s i,x, s i,y, s i,t), f(R t + S i) means we get the frame whose frame index is t + s i,x, then shift its s i,x pixels along x-axis and s i,y pixels along y-axis. The three dimensions of the image are shown in Figure 2. The original image is shown in Figures 3(a) and 3(b) and the correlation result is shown in Figure 3(c).

Figure 2.

Figure 2

Three dimensions of the original image.

Figure 3.

Figure 3

(a) An image whose frame index is t. (b) An image whose frame index is t + 1. (c) The image expression of correlation result (the bright area) of image t and image t + 1, S i = (0,5, 1).

Given a human walking cycle, images sequence consists of T frames, the features of this human gait images sequence can be described as

Ω=(Ψ(f(R1,S1),,Ψ(f(RT1),S1))Ψ(f(R1),S2),,Ψ(f(RT1),S2)Ψ(f(R1),SI),,Ψ(f(RT1),SI)). (3)

2.3. The Normalized Human Gait Image Correlation

Figure 4 shows a gait cycle. In the first stance, the person is at rest and the silhouette size is minimum, this corresponds to the valley at Figure 5; then he steps out his leg, the silhouette size will increase with his stepping out, in the sixth stand his legs and arms are completely separated and this corresponds to the maximum at Figure 5. Then, the legs and arms are closed, at stance 11 the size reaches minimum again. In fact, this is not a complete gait cycle but a “half” cycle, but the previous “half” cycle and the remaining “half” cycle are symmetrical, so in this paper we call this “half” gait cycle as one gait cycle.

Figure 4.

Figure 4

A human Gait cycle.

Figure 5.

Figure 5

Correlation image size calculated by raw features and by normalized features.

Formula (3) is the matrix expression of the curves shown in Figure 3. Ω can be used as features to identify humans. But these features are not robust. The human silhouettes are prone to be affected by the noise and deformation. We will transform the curve into standard curve to make features more robust by the following steps.

  • (1)
    Discrete Fourier transforms: by dividing the walking images into C walking cycles as shown in Figure 3, we can find the start (end) of a walking cycle by finding the minimum point in Figure 5. For every walking cycle c = (1,2,…, C), if this walking cycle consists of N c images, the value of N c may vary slightly based on different persons or different walking cycles. We can use Discrete Fourier transform to describe the features of these points in Figure 5 as follows:
    x[n]=m=0Nc1x^[m]i(2π/Nc)mn,n=0,,Nc1, (4)
    x^[m]=n=0Nc1x[n]ei(2π/Nc)mn,m=0,,Nc1. (5)
    x[n] is the discrete points and x^[m] is the Fourier coefficients.
  • (2)
    Normalization: Considering that different walking cycles have different frame count (N c), we need to “align” N c to a fixed value. In this paper, we divide every walking cycle into K parts, where K is a fixed value (in this paper K = 12). That is
    x[k]=kK(Nc1),x^[m]=k=0K1x[k]ei(2π/K)mk,m=0,,K1. (6)

Given person p, the normalized feature matrix for walking cycle c can be expressed as

Ω(p,c)=(x^[1]1,,x^[K]1x^[1]2,,x^[K]2x^[1]I,,x^[K]I). (7)

Given C walking cycles, we use the normalized Ω(p, c) as the final features. The final feature matrix for person p can be expressed as

Ω~(p)=1Cc=1CΩ(p,c). (8)

By inverse Fourier transforms (Formula (5)), we can also get the correlation image size from Ω~(p) inversely; the correlation image size calculated by Fourier transforms is shown in Figure 5.

2.4. Reduce Dimensions of Features

We define Ω- as follows:

Ω=1Pp=1PΩ~(p). (9)

P is the human count in the test dataset. That is to say, Ω- is the average value of all human gait features. Then, we calculate the covariance matrix COV as follows:

COV=(Ω~(p)Ω)(Ω~(p)Ω)T. (10)

Principal component analysis is used to reduce the dimensions of Ω~(p). The reduced Ω~(p) can be expressed as

Ω~(p)=(e1,e2,,ek)Ω~(p), (11)

where (e 1, e 2,…, e k) are the eigenvectors selected by PCA from covariance matrix COV.

2.5. Human Gait Identification

In the scenario of identification, the gait feature distance between human p and q can be defined as follows:

D(p,q)=||Ω~(p)Ω~(q)||, (12)

where || || means the Euclidean distance.

This probe p is assigned to person k by using nearest neighbor method. That is

D(p,k)=minkD(pk). (13)

In the scenario of verification, the similarity between two feature matrixes is defined as the negative of distance, that is

Sim(p,q)=D(p,q). (14)

In this paper, the similarity between a probe p and q i in the gallery is defined as z-normed similarity [27] as follows:

Sim(p,qi)=Sim(p,qi)MeaniSim(p,qi)s.d.iSim(p,qi), (15)

where s.d. is standard deviation.

FAR (False Acceptance Rate), FRR (False Rejection Rate), and EER (Equal Error Rate) are used to evaluate the performance of verification [27].

3. Experiment and Discussion

We use the CASIA database (dataset B) [28] to evaluate our method. There are total 124 people in dataset B. For every person, there are at least 3 walking sequences captured from 12 view angles. Every walking sequence consisted of 3-4 walking cycles. In this paper, most of the experiments are based on 90° view angle except with declaration. We randomly select 2 walking sequences as the training data to get the normalized human gait vector. Then we use the remained walking sequence to evaluate performance. In this paper, for the consideration of simplicity, s i,x has only three values: {−Δx, 0, −Δx}, s i,y has three values: {−Δy, 0, −Δy}, and s i,t has two values: {0, Δt}. Then, we can totally get 3 × 3 × 2 = 18S i, i = 1,…, 18. Using these shift vectors S i, we totally get 18 feature vectors for one human gait images sequence. By using PCA, we reduce these 18 vectors to 3 vectors.

3.1. Features Extracted by Image Correlation

For a human gait image sequence (consisting of about four waking cycles), the 13 features curves are shown in Figure 6  (Δx = 5, Δy = 5, Δt = 1).

Figure 6.

Figure 6

The 13 feature curves (Δx = 5, Δy = 5, Δt = 1).

3.2. The Result of Dimension Reduction

From Figure 6, we can see that the 13 image correlation curves have some similarity. That is to say that there is some redundant information. We use formula (11) to reduce the dimension. We set threshold = 90% and compress the 13 human gait feature vectors into 3 vectors. Figure 7 shows these three vectors for one human walking cycle.

Figure 7.

Figure 7

Compressed features by PCA.

We also compared the recognition accuracy result using different dimensions (that is the k in Formula (11)). The experiment result is showed in Figure 8.

Figure 8.

Figure 8

The relationship between k (dimensions) and top 1 recognition accuracy.

From Figure 8, we can see that the recognition result will reach its maximum at k = 4.

3.3. The Affection of Shift Length

From the definition of image correlation, we can see that the shift length (Δx, Δy, Δt) along these three axes affects the correlation result. The shift length along the horizontal and vertical axes should follow below principles: the features distance between two gaits images sequences calculated by these shifts should have the maximum value. That is to say, all samples should have the maximum standard deviation as follows:

Δx=argmaxΔxP=1P(Ω~(p)Ω)P1,Δy=argmaxΔyP=1P(Ω~(p)Ω)P1. (16)

To find the optimized shift length, we vary shift length along x- and y-axis from 1 to 10 pixels and calculate the standard deviation. To eliminate the affection of silhouette size, we divide the standard deviation by silhouette size. The calculation result is shown in Figure 9. The accuracy of reorganization is showed in Figure 10.

Figure 9.

Figure 9

The relationship between shift length and the standard deviation.

Figure 10.

Figure 10

The relationship between accuracy and shift length along x- and y-axis.

For completeness, we also estimate FAR (False Acceptance Rate) and FRR (False Rejection Rate) in verification mode. The ROC (Receiver Operating Characteristic) curves are shown in Figure 11.

Figure 11.

Figure 11

The ROC curves under different shift lengths along x- and y-axis.

Comparing Figures 9, 10, and 11, we can see that both the accuracy of reorganization and ERR (Equal Error Rate) will reach optimized value at approximately the maximum point of standard derivation in Figure 9. In fact, the 5 pixels length is just the moving length along x-axis between two frames and the 2 pixels length is just the average moving pixels along x-axis and y-axis between two frames.

3.4. The Affection of View Angle

As we know, the view angle may affect the recognition accuracy, so we compare the performance under differential view angle. Table 1 shows the experiment result.

Table 1.

Comparison of identification performance under different view angles.

View angle Accuracy
Rank 1 Rank 5
65% 90%
18° 51% 66%
36° 30% 51%
54° 33% 50%
72° 62% 79%
90° 82% 92%
108° 71% 88%
126° 38% 69%
144° 39% 58%
162° 44% 69%
180° 83% 89%

From Table 1, we can see that our method has better performance under view angles 0, 90, and 180. That may be because the image token from above the angle is not prone to be affected by deformation.

3.5. Discussions and Future Works

The correlation of two images in temporal domain and spatial domain produces a new image. This image contains much information about the moving object, especially the dynamic features of the human gait. In this paper, only the areas of the correlated images are is used as the features for identification. In the future, we should try to use other features about the correlation images, for example, the texture features about the correlation images, or features from projecting the correlation images to x- or y-axis. In this paper, we only talk about two images correlation. Maybe the correlation among three or more images can also be used in the human gait identification. We will also try this in the future.

4. Conclusions

With the increasing demand of security control, human gait based identification will attract more interest. In the future, gait based human identification might finally become applicable just as fingerprint and iris.

The kernel idea of this paper is based on the analysis of the image correlation. Fourier transform is used to get the features from the image correlation. We discuss the relationship between shift length and the identification accuracy. We also use CASIA gait database to validate our method. The maximum classification accuracy is nearly 90% and EER is approximately 10%.

Acknowledgment

The dataset used in this paper is provided by the Institute of Automation, the Chinese Academy of Sciences [28].

Conflict of Interests

The author declares that he has no financial and personal relationships with other people or organizations that can inappropriately influence his work. There is no professional or other personal interests of any nature or kind in any product, service, and/or company that could be construed as influencing the position presented in, or the review of, the paper entitled.

References

  • 1.Jain AK, Bolle R, Pankanti S. Biometrics: Personal Identification in Networked Society. Kluwer Academic Publishers; 1999. [Google Scholar]
  • 2.Blanke DJ, Hageman PA. Comparison of gait of young men and elderly men. Physical Therapy. 1989;69(2):144–148. doi: 10.1093/ptj/69.2.144. [DOI] [PubMed] [Google Scholar]
  • 3.Whittle MW. Clinical gait analysis: a review. Human Movement Science. 1996;15(3):369–387. [Google Scholar]
  • 4.Johansson G. Visual motion perception. Scientific American. 1975;232(6):76–88. doi: 10.1038/scientificamerican0675-76. [DOI] [PubMed] [Google Scholar]
  • 5.Stevenage SV, Nixon MS, Vince K. Visual analysis of gait as a cue to identity. Applied Cognitive Psychology. 1999;13(6):513–526. [Google Scholar]
  • 6.Wang L, Ning H, Tan T, Hu W. Fusion of static and dynamic body biometrics for gait recognition. IEEE Transactions on Circuits and Systems for Video Technology. 2004;14(2):149–158. [Google Scholar]
  • 7.Cunado D, Nixon MS, Carter JN. Automatic extraction and description of human gait models for recognition purposes. Computer Vision and Image Understanding. 2003;90(1):1–41. [Google Scholar]
  • 8.Yoo JH, Hwang D, Nixon MS. Gender classification in human gait using support vector machine. Proceedings of the 7th International Conference on Advanced Concepts for Intelligent Vision Systems (ACIVS '05); September 2005; Antwerp, Belgium. Springer; pp. 138–145. [Google Scholar]
  • 9.Lee L, Grimson WEL. Gait analysis for recognition and classification. Proceedings of the 5th IEEE International Conference on Automatic Face and Gesture Recognition (FGR '02); May 2004; IEEE; pp. 148–155. [Google Scholar]
  • 10.Yam CYC, Nixon MS, Carter JN. Automated person recognition by walking and running via model-based approaches. Pattern Recognition. 2004;37(5):1057–1072. [Google Scholar]
  • 11.Yam C, Nixon MS, Carter JN. Gait recognition by walking and running: a model-based approach. Proceedings of the 5th Asian Conference on Computer Vision; January 2002; Melbourne, Australia. [Google Scholar]
  • 12.Tafazzoli F, Safabakhsh R. Model-based human gait recognition using leg and arm movements. Engineering Applications of Artificial Intelligence. 2010;23(8):1237–1246. [Google Scholar]
  • 13.Dupuis Y, Savatier X, Vasseur P. Feature subset selection applied to model-free gait recognition. Image and Vision Computing. 2013;31(8):580–591. [Google Scholar]
  • 14.Cheng MH, Ho MF, Huang CL. Gait analysis for human identification through manifold learning and HMM. Pattern Recognition. 2008;41(8):2541–2553. [Google Scholar]
  • 15.Chen C, Liang J, Zhao H, Hu H, Tian J. Factorial HMM and parallel HMM for gait recognition. IEEE Transactions on Systems, Man and Cybernetics C. 2009;39(1):114–123. [Google Scholar]
  • 16.Yu S, Tan T, Huang K, Jia K, Wu X. A study on gait-based gender classification. IEEE Transactions on Image Processing. 2009;18(8):1905–1910. doi: 10.1109/TIP.2009.2020535. [DOI] [PubMed] [Google Scholar]
  • 17.Fan X, Li M, Ma J, Ren Y, Zhao H, Su Z. Behavior-based reputation management in P2P file-sharing networks. Journal of Computer and System Sciences. 2012;78(6):1737–1750. [Google Scholar]
  • 18.Kale A, Sundaresan A, Roychowdhury A, Chellappa R. Handbook on Pattern Recognition and Computer Vision. 2004. Gait-based human identification from a monocular video sequence. [Google Scholar]
  • 19.Liu LL, Yin Y, Qin W, Li Y. Gait recognition based on outermost contour. International Journal of Computational Intelligence Systems. 2011;4(5):1090–1099. [Google Scholar]
  • 20.Hong S, Lee H, Kim E. Probabilistic gait modelling and recognition. IET Computer Vision. 2013;7(1):56–70. [Google Scholar]
  • 21.Li Z, Shen HY, Sapra K. Leveraging social networks to combat collusion in reputation systems for peer-to-peer networks. IEEE Transactions on Computers. 2013;62(9):1745–1759. [Google Scholar]
  • 22.Hu H. Enhanced gabor feature based classification using a regularized locally tensor discriminant model for multiview gait recognition. IEEE Transactions on Circuits and Systems for Video Technology. 2013;23(7):1274–1286. [Google Scholar]
  • 23.Han J, Bhanu B. Individual recognition using gait energy image. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2006;28(2):316–322. doi: 10.1109/TPAMI.2006.38. [DOI] [PubMed] [Google Scholar]
  • 24.Chen C, Liang J, Zhao H, Hu H, Tian J. Frame difference energy image for gait recognition with incomplete silhouettes. Pattern Recognition Letters. 2009;30(11):977–984. [Google Scholar]
  • 25.Wang L, Tan T, Ning H, Hu W. Silhouette Analysis-Based Gait Recognition for Human Identification. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2003;25(12):1505–1518. [Google Scholar]
  • 26.Otsu N, Kurita T. A new scheme for practical flexible and intelligent vision systems. Proceedings of the IAPR Workshop on Computer Vision; 1988. [Google Scholar]
  • 27.Sarkar S, Phillips PJ, Liu Z, Vega IR, Grother P, Bowyer KW. The humanID gait challenge problem: data sets, performance, and analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2005;27(2):162–177. doi: 10.1109/tpami.2005.39. [DOI] [PubMed] [Google Scholar]
  • 28. CASIA Gait Database, 2009, http://www.sinobiometrics.com/

Articles from The Scientific World Journal are provided here courtesy of Wiley

RESOURCES