Abstract
Morphometrics, the quantitative analysis of shape, is used by craniofacial researchers to study abnormalities in human face shapes. Most of the work in craniofacial morphometrics uses landmark points that are manually marked on 3D face data and processed via a generalized Procrustes analysis. For large data sets this manual process is very time-consuming. Dense sets of pseudo-landmarks have also been proposed and successfully used for classification and clustering, but the two main methods in the literature are very computationally intensive. We have developed a computationally simple method that can compute pseudo-landmark points at different resolutions from 3D meshes of human faces. In this paper, we perform a comparative study employing L1-regularized logistic regression to train a classifier that predicts the sex of 500 normal adult face meshes in order to compare our method to two alternative pseudo-landmark methods and a distance matrix approach. Our results show that our method, which is fully automatic, achieved similar results to the best-scoring methods with no manual landmarking and with much lower computation time. Use of the distance matrix did not improve classification results.
I. INTRODUCTION
The FaceBase Consortium1 is a group of investigators who study craniofacial abnormalities in humans and other species and who are providing data and software tools to be shared by craniofacial researchers [6]. One project, the 3D Facial Norms Database, is studying the variation of facial forms in humans by constructing a repository of 3D facial data (in the form of 3D meshes) and genetic data. Samples of healthy Caucasian individuals are being collected in order to analyze the variation in morphology of the face. Other FaceBase projects deal with abnormalities, including cleft lip, cleft palate, and midface hypoplasia (flattening). The 3D Facial Norms Database is expected to be useful for providing control data for these and other projects studying abnormalities.
Most of the work on morphometrics in the craniofacial research community uses standard hand-marked landmarks to characterize the data. Usually, the data are aligned via these landmarks using the well-known Procrustes algorithm [8] and can then be compared using the related Procrustes distance from the mean or between individuals. For large databases, hand landmarking is very tedious, and automatically determined landmarks still have to be checked by humans. This makes the use of pseudo-landmarks an attractive alternative. There are two pseudo-landmarking methods in the medical literature: 1) the well-known dense correspondence method of Hammond and Hutton [5] and 2) the more recent anthropometric mask approach of Claes [2], both of which require an initial set of hand-marked labels to compute correspondences and have fairly heavy computation.
We have developed a very simple, but effective, method that computes pseudo-landmarks by cutting through each 3D head mesh with a set of horizontal planes and extracting a set of points from each plane. Correspondences among heads are not required, and the user does no hand marking. For classification purposes, we use L1-regularized logistic regression, which deals effectively with the high dimensionality of the data and modest number of samples. This method also allows us to easily detect the points that most contributed to any given classification task.
We compare our methodology to 1) the use of the 24 standard hand-marked landmark points, 2) the Hammond method, 3) the Claes method, and 4) the use of a distance matrix in terms of accuracy of classification and computation time. While our intended use of this methodology is for general analyses, content-based image retrieval, and grouping within the normal population, we used classification experiments for comparison because of the availability of ground truth data. Although the data came with both sex and age attributes, the task of age discrimination was very difficult, both for the computer and the human experimentors, since our data set consists of adults between 18 and 40 years of age, which are not easily discriminable based on shape alone. Thus our experiments in this paper are all on the sex attribute, which provided a very good test vehicle.
II. RELATED LITERATURE
Classic morphometrics research uses a set of standard landmarks and employs generalized Procrustes analysis [8] for comparison, clustering, or classification. Each head is represented by its configuration of landmark points, which are translated to have the same centroid and scaled to have the same size. Then, each object is rotated to best fit an arbitrarily chosen consensus head according to Procrustes distance, the Euclidean distance between the two sets of coordinates. If the total Procrustes distance for the set is small enough, the procedure terminates. Otherwise, the consensus is set to the average of the newly rotated configurations and the procedure iterates. Principal components analysis (PCA) and other methods can be applied to the dataset that is now represented in its Procrustes shape coordinates.
Hammond et al. extended classical morphometrics to their dense surface models [5] in which they obtain a correspondence between a base mesh and all other individual meshes, starting with an initial hand-marked correspondence and using Bookstein’s thin-plate spline technique [1]. Claes et al. [2] used an anthropometric mask of spatially dense, uniformly distributed points calculated on an average face from a healthy population and fit onto each other face.
Another methodolgy that used pseudo-landmarks was developed by Ruiz et al. [9] for classification of skulls with different types of craniosynostosis, a disorder in which the sutures of the skull fuse prematurely. They manually selected three CT slice planes defined by internal brain landmarks and lying parallel to the skull base plane and extracted a sequence of equally spaced points from the oriented outlines of the CT bone image at each plane. The distance matrix derived from all pairs of these points was the feature vector used in their work. Lin et al [4] derived a set of symbolic descriptors from the distance matrix by probabilistic latent semantic analysis (PSLA) [7] and used them for both classifying and clustering the data. Yang et al [13] took this approach further by fully automating the process of constructing the distance matrix from a set of planes parallel to a base plane required to pass through the nasion and opisthion points and to be perpendicular to the plane of symmetry of the head. She used different variants of logistic regression for classification.
There has also been a great deal of work in finding facial landmarks and some work on sex classification in the computer vision literature. For the most part, this work uses 2D color or gray tone photographs, rather than 3D meshes. One recent work that is worthy of mention is the classification work of Wu, Smith, and Hancock [14] which obtains a 97% accuracy on sex classification. However, their data is 2D gray tone photographs, which they go to some trouble to convert to an estimation of 3D shape from surface normal vectors. Because their representation, their data set, and their goals are completely different from our own, our results cannot be directly compared to theirs.
III. METHOD
Our method starts with 3D head meshes that have been pose-normalized to face front and whose plane of symmetry (midsagittal plane) has been computed. We compute two landmark points, the sellion and chin tip, and construct planes through these points perpendicular to the plane of symmetry of the face. Using these two planes as base planes, we construct m parallel planes through the head and from each of them sample a set of n points from the face region, where the parameters n and m are selected by the user. The set of all m × n 3D points becomes the feature vector of pseudo-landmark points representing the face, which can be used in description, quantification, and the classification experiments used for comparison results in this paper.
A. Finding Sellion and Chin Tip
Each head mesh has been pose-normalized to face straight forward and the plane of symmetry detected. The intersection of the plane of symmetry with the surface of the face is a contour. The tip of the nose is found as that point of the contour with the largest z-value. Following the contour up from the tip of the nose, the sellion is located at a local minimum for z. Following the contour down, the chin point is located at a local maximum for z. The sellion was selected as a landmark from which to begin the construction of pseudo-landmarks, because it can be reliably detected in the head meshes and is an approximate upper bound on the midface area studied by many of the researchers in the FaceBase Consortium. While the corners of the mouth would usually mark the bottom of the midface region, the tip of the chin was selected for this work, because the chin area was deemed important in sex classification.
B. Constructing Planes and Extracting Points
A plane PU through the sellion and a second plane PL through the tip of the chin, both perpendicular to the plane of symmetry of the face, are constructed. In this study, the user is asked to select how many equally-spaced y-z planes he/she would like to use between PL and PU and how many above PU, leading to the selection of m horizontal planes. The user also selects the number of points n to be extracted from the face on each plane. A horizontal slice of the head mesh is extracted at each of the m planes and restricted to the facial area by eliminating the back side of the head. Then at n equally-spaced x-positions including that of the plane of symmetry, the y and z values are sampled to obtain n 3D points on each plane. Figure 2a shows a 3D head mesh with two planes above the sellion, one at the sellion, one at the tip of the chin, and eleven in between.
Fig. 2.

(a) Head mesh showing planes at the sellion and the tip of the chin with 11 planes between and 2 planes above them. (b) Face mesh with pseudo-landmarks in 35×35 resolution. (c) Head mesh with pseudo-landmarks that were most useful for sex classification at 35×35 resolution.
C. L1-Regularized Logistic Regression
In order to compare our pseudo-landmark methodology to the other methods, we ran a large set of classification experiments, using the sex ground truth of the data. Motivated by the successful work of Yang et al [13], we used L1-regularized logistic regression [10] to learn a model to predict the sex classes of the samples in our dataset. Logistic regression learns the model parameters w0 and w in the probability function p(y|x, w) that a data sample x belongs to a certain class y.
| (1) |
Due to the high-dimensionality of the data, learning the unregularized logistic regression [4] will result in overfitting. To avoid overfitting, we applied L1 regularization that induces sparsity in the solution w such that many of the coefficients in w are set to exactly zero. L1-regularization has been rigorously proven to be effective in selecting relevant features when there are exponentially many irrelevant ones [3]. The parameters w0 and w minimize the following loss function called the lasso:
| (2) |
| (3) |
where λ is a regularization parameter for the L1-norm. L1 regularization has been widely used in many applications in which a small number of features must be selected out of many.
Additionally, when we use the original pseudo-landmark coordinates (as opposed to their principal components), the separate x, y, and z features that correspond to the same pseudo-landmark are likely to be either predictive or not predictive together. Therefore, we use a variation of L1-regularization, the group lasso [11], to induce bias such that the features corresponding to a pseudo-landmark tend to be selected together. The loss function of the group lasso is,
| (4) |
where μ is a regularization parameter for the new penalty term, and wi1, wi2 and wi3 are weights that correspond to the ith pseudo-landmark.
IV. DATA
Our head data consists of 3D meshes of 30–40,000 points obtained from several 3dMD® digital stereophotogrammetry imaging systems, which were used to capture the facial surfaces of subjects in the 3D Facial Norms Database. These systems are outfitted with multiple CCD cameras mounted at fixed angles and distances, to capture overlapping views of the face and head. The entire capture process occurs in less than 2 milliseconds and results in a dense 3D connected mesh that conforms to the geometry of the face.
Prior to 3D image capture, scalp hair obscuring the subject’s face was cleared away and subjects were positioned so that their heads were centered between the imaging pods. The heads were positioned slightly upward in order to ensure adequate coverage of the subnasal region. Meshes were cleaned to remove extraneous parts of the body and pose-normalized using a method described in [12]. Figure 1 shows a surface mesh before cleaning (a) and the cleaned and pose-normalized surface mesh produced for our analyses (b).
Fig. 1.

(a) Uncleaned mesh obtained from the 3dMD imaging system (b) Cleaned and pose-normalized version of the mesh for automated analysis (c) 20 landmark points on the cleaned mesh.
Human experts marked 24 landmarks on each mesh for use in our experiments. The 24 landmarks collected include the nasion, pronasale, subnasale, labiale superius, stomion, labiale inferius, sublabiale, ganthion, endocanthion (right), endocanthion (left), exocanthion (right), exocanthion (left), alare (right), alare (left), alar curvature point (right), alar curvature point (left), subalare (right), subalare (left), crista philtri (right), crista philtri (left), chelion (right), chelion (left), traion (right), and tragion (left). Figure 1c shows the 24 landmark points on the head of Figure 1b.
V. EXPERIMENTS AND RESULTS
Our experiments use a preliminary dataset of 500 adult heads from the 3D Facial Norms Database. We have only 3D meshes and no color data. The distribution is as follows: 250 males, 250 females, with ages between 18 and 40. In these experiments, feature extractions were run on a 64 bit Microsoft Windows 8 machine with an Intel Core i7-2600 3.40 GHz CPU and 4GB RAM, using VTK library functions for 3D processing. Classifications were run on a 64 bit Microsoft Windows Server 2008 R2 machine with Intel Xeon CPU E5649 @ 2.53 GHz (2 processors) and 32GB RAM, using MATLAB 2012b and the SLEP package.
After the processing described above, we obtain a feature vector of landmark points for each head and compute feature vectors of pseudo-landmark points at multiple different resolutions, one of which is shown in Figure 2b. We also apply the Hammond method and the Claes method to achieve feature vectors of pseudo-landmark points from those methods. Finally, we produce a distance matrix from our 35 × 35 resolution pseudo-landmarks for further comparison.
Figure 3 shows the results for sex classification using 10-fold cross validation and running each experiment 10 times with the average accuracy reported. In all experiments, 500 samples were used: 250 male and 250 female. Experiments were run both with original points (if possible) and with principal components analysis (PCA). In the PCA experiments, all principal components were used, since using less tended to reduce the accuracy.
Fig. 3.

Sex classification experiments. LMK = 24 Landmarks, P35 = 35 × 35, P45 = 45 × 45, P55 = 55 × 55, P65 = 65 × pseudo-landmarks, Ham = Hammond, Cla = Claes, Dis = 35 × 35 distance matrix, num = number, acc = accuracy. Feat (sec) = feature extraction time, PTS (sec) = classification time with points, PCA (sec) = classification time with PCA. Number of points for Hammond and Claes is approximate.
The results show that pseudo-landmarks of all categories achieved higher sex classification accuracy than the standard 24 landmarks both when the original landmark points were used for classification and when the principal components were used. All classification scores improved with PCA. All pseudo-landmark methods–our own, Hammond, and Claes–achieved about 95% accuracy using PCA, with the 35 × 35 distance matrix scoring slightly lower. Since our lowest resolution (35 × 35) feature vector worked just as well as our highest resolution (65 × 65) feature vector, it is our method of choice. The Hammond method takes an order of magnitude longer to run than either our method or the Claes method, which takes about 4 times as long as our method. Furthermore, both the Hammond method and the Claes method require a set of hand-marked landmarks for initialization of their correspondence-finding procedures, while our method is fully automatic. Using the 35 × 35 distance matrix instead of the original points did not gain any improvement in accuracy, and at higher resolutions the matrices were too large to use. Thus in sex classification, our pseudo-landmarking method, using the principal components of the original points, is superior to both hand-landmarking and the other well-known pseudo-landmarking methods.
One advantage of using logistic regression is its ability to determine the most important features (here points of the face) for a given classification task. For this work, we defined the most useful features as those for which logistic regression calculates a weight higher than an empirically determined threshold (here 0.01). Figure 2c shows the points that were most useful for sex classification at 35 × 35 resolution. As can be seen, the central areas at the level of the eyebrows are most different between adult men and women.
VI. CONCLUSIONS AND FUTURE WORK
We have described a simple, but effective methodology for extracting pseudo-landmarks from 3D craniofacial data in the form of 3D meshes. Our method, unlike the competing methods, is fully automatic; it requires no hand-landmarking by humans, making it especially suitable for large data sets. Due to the availability of ground-truth data, we have used sex classification tasks to compare our methodology to the standard 24-landmark method, the Hammond dense correspondence method, and the Claes method. All three of these methods require prior hand-landmarking. Our experiments show that our method beats the standard 24-landmark method and is equivalent to the Hammond and Claes methods in accuracy, but much faster. Use of a distance matrix in place of the original points did not improve the accuracy of classification.
The purpose of running classification experiments in this paper was to test our features for validity and then adapt them for quantifying various conditions, either abnormal or, in the case of this dataset, normal variations. The methodology developed here for pseudo-landmark extraction is general in that pseudo-landmarks can be extracted from any area of the face and then used to describe the shape in that area. A content-based retrieval system is being developed that will use this methodology, among others, to retrieve heads of similar shape from the large 3D Facial Norms Database that is being collected. The methodology will also be used to group the heads in the database according to similar facial features.
Acknowledgments
This work was supported by the National Institute of Dental and Craniofacial Research under Grant No. U01-DE020050 and Grant No. U01-DE020078.
Footnotes
References
- 1.Bookstein FL. Landmark Methods for Forms without Landmarks: Morphometrics of Group Differences in Outline Shape. /it Mecial Image Analysis. 1997;1(3):225–243. doi: 10.1016/s1361-8415(97)85012-8. [DOI] [PubMed] [Google Scholar]
- 2.Claes P, Walters M, Vandermeulen D, Clement JG. Spatially-Dense 3D Facial Asymmetry assessment in Both Typical and Disordered Growth. Journal of Anatomy. 2011;219(4):444–455. doi: 10.1111/j.1469-7580.2011.01411.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Lee S-I, Lee H, Abbeel P, Ng AY. Efficient L1 Regularized Logistic Regression. Proceedings of the 21st National Conference on Artificial Intelligence; 2006. [Google Scholar]
- 4.Lin HJ, Ruiz-Correa S, Shapiro LG, Gatica-Perez D, Hings AV, Cunningham MI, Speltz ML, Sze RW. Symbolic Shape Descriptors for Classifying Craniosynostosis Deformations From Skull Imaging. IEEE Engineering in Medicine and Biology Society. 2005:6325–6331. doi: 10.1109/IEMBS.2005.1615944. [DOI] [PubMed] [Google Scholar]
- 5.Hammond P, Hutton TJ, Allanson JE, Campbell LE, Hennekam RCM, Holden S, Patton MA, Shaw A, Temple IK, Trotter M, Murphy KC, Winter RM. 3D Analysis of Facial Morphology. American Journal of Medical Genetics. 2004;126A:339,348. doi: 10.1002/ajmg.a.20665. [DOI] [PubMed] [Google Scholar]
- 6.Hochheiser H, et al. The FaceBase Consortium: A Comprehensive Program to Facilitate Craniofacial Research. Developmental Biology. 2011;355(2):175–182. doi: 10.1016/j.ydbio.2011.02.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Hofmann T. Unsupervised learning by Probabilistic Latent Semantic Analysis. Machine Learning. 2001;42:177–196. [Google Scholar]
- 8.Rohlf F, Slice D. Extensions of the Procrustes Method for the Optimal Superimposition of Landmarks. Systems Zoology. 39:40–59. [Google Scholar]
- 9.Ruiz-Correa S, Sze RW, Lin HJ, Shapiro LG, Speltz ML, Cunningham ML. Classifying Craniosynostosis Deformations by Skull Shape Imaging. IEEE Computer-Based Medical Systems. 2005:335–340. doi: 10.1109/IEMBS.2005.1615944. [DOI] [PubMed] [Google Scholar]
- 10.Tibshirani R. Regression Shrinkage and Selection Via the Lasso. Journal of the Royal Statistical Society B. 1994;58:267–288. [Google Scholar]
- 11.Yuan M, Lin Y. Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society B, Methodological. 2006;68:49–67. [Google Scholar]
- 12.Wilamowska K, Shapiro LG, Heike CL. Classification of 3D Face Shape in 22q11.2 Deletion Syndrome. International Symposium on Biomedical Imaging; 2009. p. 537?537. [Google Scholar]
- 13.Yang S, Shapiro LG, Cunningham ML, Speltz M, Lee S-I. Classification and Feature Selection for Craniosynostosis. ACM Conference on Bioinformatics, Computational Biology and Biomedicine; 2011. pp. 340–344. [Google Scholar]
- 14.Wu J, Smith WAP, Hancock ER. Gender Discrimination Models from Facial Surface Normals. Pattern Recognition. 2011;44:2871–2886. [Google Scholar]
