Abstract
Measuring food volume (portion size) is a critical component in both clinical and research dietary studies. With the wide availability of cell phones and other camera-ready mobile devices, food pictures can be taken, stored or transmitted easily to form an image based dietary record. Although this record enables a more accurate dietary recall, a digital image of food usually cannot be used to estimate portion size directly due to the lack of information about the scale and orientation of the food within the image. The objective of this study is to investigate two novel approaches to provide the missing information, enabling food volume estimation from a single image. Both approaches are based on an elliptical reference pattern, such as the image of a circular pattern (e.g., circular plate) or a projected elliptical spotlight. Using this reference pattern and image processing techniques, the location and orientation of food objects and their volumes are calculated. Experiments were performed to validate our methods using a variety of objects, including regularly shaped objects and food samples.
Keywords: Dietary assessment, Volume estimation, Circular referent, Image processing
1. Introduction
Accurate dietary assessment is an important procedure in both clinical and research studies of many diseases and disorders, such as obesity and diabetes (Franz et al., 2002; Mokdad et al., 2004; Knowler et al., 2009; Thompson and Subar, 2001). However, a major problem exists: the assessment is often inaccurate for individuals in a real life setting (Thompson and Subar, 2001; Wacholder, 1995; Trabulsi and Schoeller, 2001; Johansson et al., 1998). The self-reporting of food intake (e.g., food record, 24-h dietary recall and food frequency questionnaire) has been the primary method for acquiring information about consumed food and the portion size (Thompson and Subar, 2001; Wacholder, 1995; Trabulsi and Schoeller, 2001; Johansson et al., 1998; Goris and Westerterp, 1999; Goris et al., 2000; Most et al., 2003; Pietinen et al.,1988a,b; Chu et al., 1984; Martin et al., 2007). However, self-report methods are known to be biased, due to significant underreporting (Trabulsi and Schoeller, 2001; Johansson et al., 1998; Goris and Westerterp, 1999; Goris et al., 2000). In order to improve assessment accuracy, many portion-size measurement aids (PSMAs) have been developed (Posner et al., 1992; Faggiano et al., 1992; Kuehneman et al., 1994; Nelson et al., 1994; Lucas et al., 1995; Cypel et al., 1997; Godwin et al., 2006; Nelson and Haraldsdottir, 1998; Frobisher and Maxwell, 2003; Williamson et al., 2003; Turconi et al., 2005; Ershow et al., 2007; Foster et al., 2005; Martin et al., 2009). Common PSMAs include household measures (e.g., rulers), food models (real food samples or abstract models), and food photographs. While such PSMAs are useful in estimating portion size, they provide only a rough volumetric measure and are not always suitable or convenient in practice. Recently, camera-enabled cell phones and other mobile devices (e.g., iPad) have become very popular. These devices can be used to take food pictures conveniently, which can then be remotely transmitted to dietitians for dietary analysis (Kikunaga et al., 2007; Boushey et al., 2009; Zhu et al., 2010; Ngo et al., 2009).
Although digital images provide an excellent tool to document foods consumed, it is difficult to estimate food portion size from a single image without a dimensional referent, except in special cases (Du and Sun, 2006; Wang and Nguang, 2007). Accordingly, several approaches have been studied that include a dimensional referent in the field of view, such as a checkerboard (Boushey et al., 2009; Zhu et al., 2010; Subar et al., 2010; Sun et al., 2008, 2010a,b; Jia et al., 2009; Yue et al., 2010). However, it can be inconvenient and annoying for subjects constantly to have to remember to carry a checkerboard, and to place it beside the food when eating (Boushey et al., 2009; Zhu et al., 2010; Subar et al., 2010). Further, eating behavior changes when subjects are asked to document food intake. For example, they may reduce intake significantly, an effect frequently observed in obese subjects (Goris and Westerterp, 1999; Goris et al., 2000). In order to obtain objective data on food intake, we have developed a low-cost electronic device containing a miniature camera and a number of other sensors to record food intake automatically (Sun et al., 2008, 2009, 2010a,b). This device is designed to be almost completely passive to the subject, and thus hopefully not intrude on or alter the subject’s eating activities. This button-like device can be worn using a lanyard around the neck or attached to clothes using a pin or a pair of magnets (see Fig. 1). When utilized, it takes pictures of the scene in front of the subject at a pre-set rate (e.g., 2–4 s/frame). All the images are stored on a memory card. After the dietary study is complete (e.g., 1 week), the data are transferred to the dietitian’s computer for further analysis. Since images of the entire eating period are recorded (see Fig. 2), a pair of images can always be selected to represent each food item before and after eating (Fig. 2(a) and (d)). Once each food item is identified by the dietitian and its consumed volume is determined from the “before” and “after” images, our software system calculates calorie and nutrient values for each food using data available from the United States Department of Agriculture (USDA) Food and Nutrient Database for Dietary Studies (FNDDS) database (USDA, 2010). The energy intake during the study period can be accumulated and a detailed report on calorie and nutrient intake created for dietary assessment.
Fig. 1.
(a) Subject wearing the device; (b) electronic circuit board within the device.
Fig. 2.
Typical pictures before (a), during (b–c) and after eating (d).
In FNDDS, the conversion between measurable/observable volumes (e.g., cup, cubic inch, piece, slice, large/medium/small) and weight is available for most foods, such as cheeses, breads, vegetables and meats. For example, the weights of one cubic inch cheese and a cup of stewed pork chop are 17.3 and 134 g, respectively. Therefore the energy and nutrient contents of a specific food can be estimated based on its volume, as calculated from an image. In the current version of the FNDDS database, certain foods are listed without volumetric measures. In order to solve this problem, several research groups, including us, have been collaborating with USDA to conduct research on food density measures.
In the current study, we focus on image-based food volume estimation and present two novel methods for calculating food volume from a single image. Previous studies have shown that, under certain conditions, the elliptical image of a known circular object can provide sufficient information about the location and orientation of any planar object (Shin and Ahmad, 1989; Safaee-Rad et al., 1992). In our study, we have proposed two types of approaches based on the elliptical patterns extracted from a single input image (Jia et al., 2009; Yue et al., 2010). The first type is based on a circular object (e.g., a circular plate) which is commonly present during eating. We call the corresponding portion size estimation method the “plate method”. The second type is based on a conic spotlight projected by a light emitting diode (LED) installed on our wearable electronic device. We call this method the “LED method”. In either method, we first determine the location and orientation of an object plane by interactive elliptical pattern extraction and computation using our algorithms. Then, dimensional variables (e.g., length and thickness) are measured by selecting several feature points on the food image. The measured dimensions provide an estimate of food volume. In the following, we will describe our methods in detail and discuss our experimental results.
2. Methods
We model the camera using the pinhole model which is commonly utilized in photography (Sonka et al., 2007; Hartley and Zisserman, 2003). For simplicity, we assume that the camera has a fixed focal length, which is true in our wearable device and most cell phones; and that the physical image plane of the camera is perpendicular to its optical axis, which is almost universally true. Under these assumptions, the intrinsic parameters of the camera (e.g., the focal length) can be obtained either from manufacturer’s specifications or using a standard calibration procedure (Tsai, 1987). Based on the intrinsic parameters, the indices of each pixel in the digital image can be converted to the real-world coordinates in the image plane (Hartley and Zisserman, 2003; Sonka et al., 2007). For simplicity, we use the term “image” to refer to the image on the physical image plane. According to these assumptions, the relationship between the object plane (e.g., a table surface) and the image plane can be obtained by means of a perspective projection model. We describe this model with respect to the plate and LED methods separately in the following two sections.
2.1. Object location and orientation using the plate method
We first assume that the food item to be estimated is on an object plane in front of the subject, as in most dining environments. We further assume that the camera can be posed arbitrarily towards the food. The relationship between a circular object (e.g., dinning plate) on the object plane and its image is illustrated in Fig. 3(a).
Fig. 3.
Geometric model of the system and schematic diagrams of cone transformation. (a) Perspective relationship between circular object and its image: Origin O is the optical center of the camera and xyz defines the real word coordinate system where z is perpendicular to the image plane. (b) New coordinate system XYZ after applying cone transformation T: in this system, the cone is centralized. (c) New coordinate system X′Y′Z′ after applying another cone transformation: in this system, the object plane is parallel to the X′OY′ plane.
In this model, the image of a circular object becomes an ellipse/circle in the image plane under the pinhole model assumption. The equation of the ellipse is given by
(1) |
Eq. (1) is known after fitting the boundary of the circle in the digital image. This boundary defines the base of a quadric cone. The equation of the cone in the three-dimensional coordinate system defined in Fig. 3(a) can be formulated in the following quadratic surface function (Safaee-Rad et al., 1992):
(2) |
where f is the focal length of the camera.
The object plane can be defined as:
(3) |
where (l, m, n) represents the norm of the plane with l2 + m2 + n2 = 1. The intersection between the quadric cone and the object plane represents the boundary of the circular object. In order to simplify the expressions for (l, m, n), two cone transformations are utilized to centralize the quadric cone and then make the object plane parallel to the X′O′Y′ plane, see Fig. 3(b) and (c). After applying these transformations, the values of l, m, and n can be determined analytically (Safaee-Rad et al., 1992). Although there are four symmetrical solutions with respect to the origin of XYZ frame (Fig. 3(b)), it can be uniquely determined in our practical case since the wearable camera always takes pictures when the wearer is beside the food.
Under these given conditions, it has been shown (Safaee-Rad et al., 1992) that the solutions of l, m, n are:
(4) |
where L = 0, , λ1, λ2, λ3 are the eigen values of matrix are the three eigen vectors corresponding to the eigen values λi of Q in Eq. (3), i = 1, 2, 3.
Although the orientation of the object has been obtained, its location is still non-unique because parallel intersection planes of a conicoid are all similar (Bell, 1944), producing the same image on the image plane. However, if the size of the circular object is known, its location can be uniquely determined according to the geometry of perspective projection. For convenience, we apply another coordinate transformation, the equation of the intersection plane in the new coordinate system X′Y′Z′ becomes Z′ = p, see Fig. 3(c). Then, the vertical distance p between the origin O and the object plane is given by (Safaee-Rad et al., 1992):
(5) |
here r is the radius of the plate which must be provided, and C1–C4 are parameters calculated from the cone transformation:
(6) |
2.2. Object location and orientation using the LED method
In this method, a narrow beam LED is positioned at the wearable device in a small distance from the camera. When the picture is taken, the LED produces a 16° conic spotlight into the field of view. The LED light can be either visible or invisible to the human eye, but must be visible to the camera. The LED light is selected carefully so that the spotlight in the cross-section perpendicular to the axis of the beam is as circular as possible.
In order to simplify the mathematic model of the measurement system using the LED as a referent, we assume that the optical axes of the camera and the light beam are in the same plane and parallel to each other. This assumption can be satisfied by carefully calibrating the wearable device. The resulting model is shown in Fig. 4. O and D are the optical centers of the camera lens and the LED, respectively, and β is the falloff angle of the LED cone. The spotlight is on the intersection plane between the camera and the LED cones.
Fig. 4.
Geometric model of measurement system using an LED: O and D are optical centers of the camera lens and LED, respectively, and β is the falloff angle of the LED cone. The elliptical spotlight is on the intersection plane (object plane) between the camera cone and the LED cones.
In this model, the equation of the camera cone in the coordinate system xyz can be derived in a similar way as that in Eq. (2), which can be written in the matrix form:
(7) |
The norm of and z-distance of the object plane are, respectively, (l, m, n) and d0 (see Fig. 4). Because (0, 0, d0) is on the object plane, the plane equation is given by
(8) |
(9) |
In order to ease mathematical treatment, the following rotational transformation is applied (Safaee-Rad et al., 1992)
(10) |
If l = m = 0, the object plane is perpendicular to the optical axis of the camera. In this case, it can be derived easily that the image of the spotlight is still a circle and the distance d0 can be uniquely determined by the position of the circle in the image. The further the distance is, the closer the circle is to the center of the image.
Substituting (10) into (7) yields
(11) |
Under the new x′y′z′ coordinate systems, the object plane equation becomes
(12) |
Substituting Eq. (12) into Eq. (11), we can obtain the equation of the elliptical spotlight on the object plane as the intersection between the object plane and the camera cone:
(13) |
Eq. (13) can be written in the form of a quadratic equation (detailed coefficients are listed in Appendix A):
(14) |
Besides the camera cone, there is another cone, called LED cone, in this model. The equation of the LED cone in the original coordinate system xyz can be derived easily:
(15) |
Following the same procedure as Eqs. (7)–(13), the equation describing the intersection between the LED cone and the object plane with respect to the x′y′z′ coordinate system is given by
(16) |
This equation can be written as (see Appendix A for details)
(17) |
Since (14) and (17) represent the same ellipse on the object plane (see Fig. 4), the coefficients in these two equations should be the same, resulting in five equations, see Appendix A. We can solve these equations to obtain the information about the location and orientation of the object plane through iterative methods (Martinez, 1991) or optimization methods (Grosan and Abraham, 2008). However, it is generally very hard to obtain an analytical solution. Here we only consider a particular case where the plate is placed on a leveled surface, which is simple but common in our application. Let the relative rotation angle between camera and table be only in one direction. Under these assumptions, we define the sharp angle between the object plane and the optical axis of the camera to be θ. Then, the equations can be simplified and solved as (Appendix A):
(18) |
(19) |
The normalized orientation vector of the object plane is given by:
(20) |
And the distance p between the origin O and the object plane is
(21) |
2.3. Food volume estimation
In order to estimate the volume of food which can be determined using several dimensional measurements (e.g., length, width, height, and/or diameter), we must derive expressions for the distance of two arbitrary points that are either parallel to the object plane or perpendicular to it.
Fig. 5 shows the scheme of calculating the length DE which is on the object plane and the height FE which is perpendicular to the object plane. In Fig. 5(a), line AB is the perspective projection of DE in the observed image. We then define BC as a supplementary line which is parallel to DE in the plane ODE. For convenience, we represent A and B in the transformed X′Y′Z′ coordinate system specified by Eq. (10). Since, in this system, point C has the same coordinate value as the point B in the Z′-direction, and C is on the same line of OA whose equation is known, so |BC| can be calculated. With the known value of |BC|, length |DE| can be calculated from represents the Z′ coordinate of points B and C, and p is given by Eq. (5) or Eq. (21).
Fig. 5.
Schematic representation of length (left panel) and height (right panel) measurements.
The height FE can be derived in a very similar way, as shown in Fig. 5(b). Line EF is perpendicular to the plate plane with points E and F being on and off the plane, respectively. Let line HG be perspective projection of FE in the observed image. Draw a line GI parallel to EF and intersects line OF at point I. Since point H, G and line OH are known, |GI| is determined. Because ΔOIG and ΔOFE are similar, represent the Z coordinate of point G under X′Y′Z′. Thus, length |EF| is fully determined.
3. Experimental results
3.1. Experiment setup
We have conducted experiments to validate the feasibility of our food volume measurement methods. Four regularly shaped objects and six food replicas (Nasco, Fort Atkinson, Wisconsin) were utilized in our experiments (Fig. 6). The regularly shaped objects included a LEGO block, a small clip box, a large clip box and a golf ball. The food replicas included a piece of cornbread, a thin slice of onion, a whole hamburger, a glass of juice, a peach and an orange which were all realistically shaped and colored. These replicas represent three common shapes – prism, cylinder, cone and sphere. The ground truth for regular-shaped objects was measured by a micrometer. Since the volume of food replica provided by the manufacture is not accurate, the true volumes were the average over three measurements using the water displacement method. Each food replica was then modeled by a simple geometry shape, e.g., the cornbread was modeled by a prism with squared bottom, and the hamburger/onion was modeled by a cylinder. Under this assumption, the height of the food replica, except onion, was measured by a micrometer. The other dimensional information (length/radius) was then derived from the geometry model. For onion, the radius was measured while the height was derived because the top surface of the onion was not flat and not exactly parallel to the bottom. The radius peach/orange is calculated from the volume assuming its shape is a perfect sphere.
Fig. 6.
Samples used in experiment.
In order to achieve comparable data, our prototype device with a camera and an LED referent was fixed above a table and faced toward the sample (e.g., a block) on the table. Each sample was placed in four fixed locations on the table and one picture was taken for each location (see Fig. 7 for typical pictures). In this experiment, since the camera was fixed during the experiment, the LED spotlight was projected only once. In order to obtain a clear spotlight, white copy paper was placed on the table. Then, the location and orientation of the table was calculated and utilized to compute the volume of each sample. Circles printed on a paper were utilized as plates, making it easier to place different samples in the same location. The deformation of the big circles (16 cm in diameter) around the sample was utilized to estimate the food volume.
Fig. 7.
Each object (here a piece of cornbread) placed at four different locations (L1–L4) where the circular patterns represent hypothetical plates.
3.2. Determination of the object’s location and orientation
In our algorithm, the fitted parameter of the elliptical feature determines the location and orientation of the object plane. Currently these parameters are estimated by fitting eight manually-selected points on the border of the circular pattern or LED light pattern. The inaccuracy in the selection of points causes a certain error in the estimation of location and orientation. Since the boundary of the LED light pattern has a diffused region, ellipse fitting is not exact (see the left panel in Fig. 8). Thus, the variance in the estimated location and orientation from the 10 fitting results of the boundary points is larger for the LED method (see Table 1). In particular, the measured variance in the estimated angles was 1.8° for the LED method, while less than 0.3° for the plate method. Therefore, in the LED method, we used averaged location and orientation to compute food volume. In the plate method, only one fitting was required because the boundary was clear.
Fig. 8.
Ten fitting results for the LED spotlight (a) and plate (b). Only part of the image with the reference is shown.
Table 1.
Averaged location and orientation estimates of the object plane from 10 fitting results for the LED spotlight and circle patterns at four different locations.
LED | Plate (L1) | Plate (L2) | Plate (L3) | Plate (L4) | |
---|---|---|---|---|---|
Location (mm) | 434.0 ± 4.3 | −26.2 ± 3.1 | 42.3 ± 2.2 | 8.8 ± 2.2 | −5.1 ± 1.7 |
157.3 ± 0.6 | −159.3 ± 1.3 | −15.1 ± 0.8 | −14.1 ± 0.7 | ||
556.4 ± 1.8 | 572.2 ± 2.7 | 428.1 ± 1.2 | 409.1 ± 1.5 | ||
Orientation (°) | 46.24 ± 1.80 | 0.90 ± 0.25 | 0.61 ± 0.18 | 0.35 ± 0.25 | 0.07 ± 0.23 |
40.30 ± 0.27 | 40.16 ± 0.19 | 40.04 ± 0.13 | 40.78 ± 0.23 | ||
40.7 ± 0.26 | 40.86 ± 0.19 | 41.00 ± 0.13 | 40.24 ± 0.23 |
3.3. Results of volume estimation
From each picture, we estimated the volume of the object using the following steps. First, a number of points were selected on the observed boundary of the circular pattern. Then, the best-fit ellipse was computed using the least-square algorithm based on the selected points to determine the location and orientation of the circular pattern. Next, a food model was selected for the object (e.g., a sphere model for an orange or a prism model for a piece of cornbread) and a number of point pairs that measure the length, width, height, and/or diameter of the object in the picture were manually selected at appropriate locations. Then, these dimensional measures in pixel units were automatically converted to real-world measures in the millimeter unit. Finally, the volume of the object was computed using the location and orientation calculated either from the LED reference or from the circular reference. Following this procedure, we were able to compare the difference between two references because all other factors were constant.
In order to quantify the measurement error, the Root Mean Square Error (RMSE) was used in this study. The RMSE was defined as, where V̂j is the estimated value, V is the ground truth and n is the number of images. The volumetric estimates of all test objects are shown in Fig. 9. The averaged RMSE values of the volumetric measures of all 10 samples were 12.01% and 29.01% for plate and LED methods, respectively. It can be seen that, in general, the estimated volume using the plate method was more accurate than the LED method.
Fig. 9.
Comparison of volume estimation using circular plate and LED methods.
The estimated results of dimensions (length, width, height, radius) are illustrated in Fig. 10. Most of the RMSE values in the dimensional estimates were below 10%. Larger errors were mainly in the height or diameter estimates. These indicated a practical difficulty since, usually, the height or diameter/radius cannot be well defined for foods without corners/edges or with only round edges, such as a hamburger, a peach or an orange. In the extreme case of onion slice, the estimated errors of the radius and height were both above 10%, but the error in volume estimation using the plate method was only 4.3%. The reason was found to be the over-estimation in radius and under-estimation in height.
Fig. 10.
Comparison of dimensional estimation using circular plate and LED methods.
In order to further improve estimation accuracy, it became necessary to explore the factors affecting the measurement results. In the following, we investigated two major factors using the plate method.
3.3.1. The sensitivity of volume estimation to object size
The picture of a LEGO block was further studied to test whether the estimation of a larger object leads to a larger error. There were eight stacks in this block, representing eight objects with different number of stacks (see Fig. 6). Fig. 11(a) indicates that the absolute measurement error increases as the size of the object increases, but this trend disappeared when divided by its true volume, i.e., when the relative error was compared (Fig. 11(b)). This suggested that the measurement error was independent of food size.
Fig. 11.
Volume estimation of a LEGO block using circular plate and LED methods.
3.3.2. Reliability of selecting paired points in the image
In order to study the error caused by paired point selection in the object plane, we performed an experiment where the ellipse parameters were kept constant. Two pairs of points on the white paper, representing the diameters of the small and big circles in Fig. 8(b) (80 and 160 mm, respectively), were selected from the image. The same points were selected 10 times. The length estimation results are shown in Fig. 12. The standard deviation of the relative error over the 10 estimations was below 1%. These results indicated that point pair selection did not affect the estimation accuracy significantly.
Fig. 12.
Stability and accuracy in length estimation.
4. Conclusions and discussion
In practical cases where our wearable device is utilized, a circular plate with known diameter may not always be available, such as in fast food restaurants. In contrast, the spotlight is always present in the image, since it is always projected by the wearable device. But the projected spotlight may be liable to distortion by the surface of the food or its container. Therefore, in this study we proposed two approaches, based on an elliptical pattern produced by a dining plate or an LED beam. The feasibility of these approaches has been examined experimentally using 10 test objects and food samples. The average error was 12.01% for the plate method and 29.01% for the LED method. These results demonstrate that food volume can be determined computationally from a single image, as long as one of these referents is present in the image. Further work is needed, however, to improve the accuracy of size estimations.
In our experiments, an ultra-bright LED was used to project the spotlight. Due to the diffraction effect when the light is projected onto a surface, the border of the ellipse observed in the image is blurred. The brightness of ambient light also affects the identification of the spotlight. As a result, errors in fitting the border of the spotlight in an image occur, and may result in a larger error in volume estimation using the LED method. Other types of structured lights, e.g., a circular laser beam, may be investigated in the future.
From the experimental results, we can see that the estimated volume is more accurate for the prism-like objects than cylindrical or spherical objects. This may be caused by the difficulty in selecting the pairs of points to represent the object’s height or diameter. We know that a circular object generally becomes an ellipse in the image unless the optical axis of the camera is perpendicular to the object, in which case the object image is circular. It is, therefore, usually difficult to select two points representing the true diameter of a circular/sphere object in its observed image. The difficulty of selecting points to represent an object’s height is also a problem, most notably for irregularly-shaped objects. The larger error in estimating the height of a hamburger was mainly due to this type of measurement ambiguity.
Occlusion is another problem in the current approach. Since we assume that the bottom of the food and the reference are in the same object plane, the bottom is known when the orientation and location of the object plane is determined. However, if the bottom is occluded, we have to assume that the upper surface of the model (e.g., prism) is parallel to its bottom by a distance (i.e., height). So the height of the object has to be estimated first in order to determine the plane that the upper surface locates. As a result, the estimation error will be even larger since the error in the height estimation will influence the dimensional estimation in the upper surface.
Because many foods are not regularly shaped and the ambiguity and occlusion in dimensional estimation are common in food pictures, alternative methods that do not use, or are less dependent on, the length measurement are desirable and warrant future investigation.
Although the two approaches proposed in this study are useful in estimating food volume from one single picture, a reference object is still necessary when applying either approach. Studies are being performed in our laboratory to remove this constraint. In addition, we are also conducting research on further improvement of the accuracy in food volume estimation by fitting food items with more flexible shapes. Finally, we would like to point out that food volume estimation is only one part of the system to determine food calories and nutritional contents. The accuracy of the volumetric and nutritional information in the food database also plays an essential role in the system. We expect the accuracy of the food database to improve continuously as the food volume measurement techniques advance.
Besides calculating energy intake from the food images, our system also estimates energy expenditure by integrating the multimedia data acquired by several sensors available in our wearable device (e.g., camera, GPS, light sensor, and accelerometer). We identify human activities (e.g., eating, walking, driving) and their durations from the automatically saved time stamps. A special physical activity database (Ainsworth et al., 2000) is then utilized to estimate the caloric expenditure. Thus, our device can provide a convenient, effective measure of energy balance between energy intake and expenditure, which is highly desirable in the study of obesity and related conditions.
Acknowledgment
This work was supported by National Institutes of Health grant U01 HL91736.
Appendix A. Solution of the LED model
The intersection curve between the camera cone and the object plane can be obtained as:
(A1) |
where:
And the intersection curve between the LED cone and the object plane is:
(A2) |
where
Since (A1) and (A2) and represent the same elliptical spotlight on the object plane, the coefficients in these two equations should be the same:
(A3) |
Using (A3) and (9), we can solve for the location and orientation of the object plane. However, it is generally hard (if possible at all) to obtain a close-form solution of (A3) and (9). We have to rely on numerical approaches to solve these nonlinear simultaneous equations, e.g., through iterative methods (Martinez, 1991) or optimization methods (Grosan and Abraham, 2008). On the other hand, we can concentrate on a special case, which is simple but most commonly seen in our application: the dining table is horizontal, and the camera rotates in only one direction relative to the table. In this case, m = 0, and the equations can be simplified as:
(A4) |
where . If we define the sharp angle between the object plane and the optical axis of the camera as θ, , then t = tgθ.
(A5) |
(A6) |
In our application, d0 > 0, d > 0, and b′/a′ > 0. From (A6), we get:
(A7) |
From (A5), we get:
(A8) |
Compare (A7) and (A8), only negative symbol is feasible in our application.
Therefore,
(A9) |
References
- Ainsworth BE, Haskell WL, Whitt MC, Irwin ML, Swartz AM, Strath SJ, O’Brien WL, Bassett DR, Jr, Schmitz KH, Emplaincourt PO, Jacobs DR, Jr, Leon AS. Compendium of physical activities: an update of activity codes and MET intensities. Medicine and Science in Sports and Exercise. 2000;32(Suppl. 9):S498–S504. doi: 10.1097/00005768-200009001-00009. [DOI] [PubMed] [Google Scholar]
- Bell RJT. An Elementary Treatise on Coordinated Geometry of Three Dimensions. third ed. London, UK: Macmillan; 1944. [Google Scholar]
- Boushey CJ, Kerr DA, Wright J, Lutes KD, Ebert DS, Delp EJ. Use of technology in children’s dietary assessment. European Journal of Clinical Nutrition. 2009;63(Suppl. 1):S50–S57. doi: 10.1038/ejcn.2008.65. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chu S, Kolonel LN, Hankin JH, Lee J. A comparison of frequency and quantitative dietary methods for epidemiologic studies of diet and disease. American Journal of Epidemiology. 1984;199:323–334. doi: 10.1093/oxfordjournals.aje.a113751. [DOI] [PubMed] [Google Scholar]
- Cypel YS, Guenther PM, Petot GJ. Validity of portion-size measurement aids: a review. Journal of the American Dietetic Association. 1997;97:289–292. doi: 10.1016/S0002-8223(97)00074-6. [DOI] [PubMed] [Google Scholar]
- Du C-J, Sun D-W. Estimating the surface area and volume of ellipsoidal ham using computer vision. Journal of Food Engineering. 2006;73(3):260–268. [Google Scholar]
- Ershow AG, Ortega A, Baldwin JT, Hill JO. Engineering approaches to energy balance and obesity: opportunities for novel collaborations and research. Journal of Diabetes Science and Technology. 2007;1(1):95–105. doi: 10.1177/193229680700100115. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Faggiano F, Vineis P, Cravanzola D, Pisani P, Xompero G, Riboli E, Kaaks R. Validation of a method for the estimation of food portion size. Epidemiology. 1992;3:379–382. doi: 10.1097/00001648-199207000-00015. [DOI] [PubMed] [Google Scholar]
- Foster E, Matthews JNS, Nelson M, Harris JM, Mathers JC, Adamson AJ. Accuracy of estimates of food portion size using food photographs – the importance of using age-appropriate tools. Public Health Nutrition. 2005;9(4):509–514. doi: 10.1079/phn2005872. [DOI] [PubMed] [Google Scholar]
- Franz MJ, Bantle JP, Beebe CA, Brunzell JD, Chiasson JL, Garg A, Holzmeister LA, Hoogwerf B, Mayer-Davis E, Mooradian AD, Purnell JQ, Wheeler M. Evidence-based nutrition principles and recommendations for the treatment and prevention of diabetes and related complications. Diabetes Care. 2002;25(1):202–212. doi: 10.2337/diacare.25.1.148. [DOI] [PubMed] [Google Scholar]
- Frobisher C, Maxwell SM. The estimation of food portion sizes: a comparison between using descriptions of portion sizes and a photographic food atlas by children and adults. Journal of Human Nutrition and Dietetics. 2003;16(3):181–188. doi: 10.1046/j.1365-277x.2003.00434.x. [DOI] [PubMed] [Google Scholar]
- Godwin S, Chambers Iv E, Cleveland L, Ingwersen L. A new portion size estimation aid for wedge-shaped foods. Journal of the American Dietetic Association. 2006;106(8):1246–1250. doi: 10.1016/j.jada.2006.05.006. [DOI] [PubMed] [Google Scholar]
- Goris AH, Westerterp-Plantenga MS, Westerterp KR. Undereating and under recording of habitual food intake in obese men: selective underreporting of fat intake. The American Journal of Clinical Nutrition. 2000;71(1):130–134. doi: 10.1093/ajcn/71.1.130. [DOI] [PubMed] [Google Scholar]
- Goris AH, Westerterp KR. Underreporting of habitual food intake is explained by undereating in highly motivated lean women. The Journal of Nutrition. 1999;129(4):878–882. doi: 10.1093/jn/129.4.878. [DOI] [PubMed] [Google Scholar]
- Grosan C, Abraham A. A new approach for solving nonlinear equations systems. IEEE Transactions on Systems, Man, and Cybernetics, Part A. 2008;38:698–714. [Google Scholar]
- Hartley R, Zisserman A. Multiple View Geometry in Computer Vision. second ed. Cambridge, UK: Cambridge University Press; 2003. [Google Scholar]
- Jia W, Zhao R, Fernstrom JD, Fernstrom MH, Sclabassi RJ, Sun M. A food portion size measurement system for image-based dietary assessment. Proceedings of the 35th Northeast Biomedical Engineering Conference; Cambridge, MA, USA. 2009. [Google Scholar]
- Johansson G, Akesson A, Berglund M, Nermell B, Vahter M. Validation with biological markers for food intake of a dietary assessment method used by Swedish women with three different dietary preferences. Public Health Nutrition. 1998;1(3):199–206. doi: 10.1079/phn19980031. [DOI] [PubMed] [Google Scholar]
- Kikunaga S, Tin T, Ishibashi G, Wang DH, Kira S. The application of a handheld personal digital assistant with camera and mobile phone card to the general population in a dietary survey. Journal of Nutritional Science and Vitaminology. 2007;53(2):109–116. doi: 10.3177/jnsv.53.109. [DOI] [PubMed] [Google Scholar]
- Knowler WC, Fowler SE, Hamman RF, Christophi CA, Hoffman HJ, Brenneman AT, Brown-Friday JO, Goldberg R, Venditti E, Nathan DM. 10-year follow-up of diabetes incidence and weight loss in the Diabetes Prevention Program Outcomes Study. Lancet. 2009;374(9702):1677–1686. doi: 10.1016/S0140-6736(09)61457-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kuehneman T, Stanek K, Eskridge K, Angle C. Comparability of four methods for estimating portion sizes during a food frequency interview with caregivers of young children. Journal of the American Dietetic Association. 1994;94(5):548–551. doi: 10.1016/0002-8223(94)90222-4. [DOI] [PubMed] [Google Scholar]
- Lucas F, Niravong M, Villeminot S, Kaaks R, Clavel-Chapelon F. Estimation of food portion size using photographs: validity, strengths, weaknesses and recommendations. Journal of Human Nutrition and Dietetics. 1995;8(1):65–74. [Google Scholar]
- Martin CK, Anton SD, York-Crowe E, Heilbronn LK, VanSkiver C, Redman LM, Greenway FL, Ravussin E, Williamson DA for the Pennington CALERIE Team. Empirical evaluation of the ability to learn a calorie counting system and estimate portion size and food intake. British Journal of Nutrition. 2007;98(2):439–444. doi: 10.1017/S0007114507708802. [DOI] [PubMed] [Google Scholar]
- Martin CK, Han H, Coulon SM, Allen HR, Champagne CM, Anton SD. A novel method to remotely measure food intake of free-living people in real-time: the Remote Food Photography Method (RFPM) British Journal of Nutrition. 2009;101(3):446–456. doi: 10.1017/S0007114508027438. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Martinez JM. Quasi-Newton methods for solving underdetermined nonlinear simultaneous equations. Journal of Computational and Applied Mathematics. 1991;34(2):171–190. [Google Scholar]
- Mokdad AH, Marks JS, Stroup DF, Gerberding JL. Actual causes of death in the United States, 2000. The Journal of American Medical Association. 2004;291(10):1238–1245. doi: 10.1001/jama.291.10.1238. [DOI] [PubMed] [Google Scholar]
- Most MM, Ershow AG, Clevidence BA. An overview of methodologies, proficiencies, and training resources for controlled feeding studies. Journal of the American Dietetic Association. 2003;103(6):729–735. doi: 10.1053/jada.2003.50132. [DOI] [PubMed] [Google Scholar]
- Nelson M, Atkinson M, Darbyshire S. Food photography. I: The perception of food portion size from photographs. British Journal of Nutrition. 1994;72(5):649–663. doi: 10.1079/bjn19940069. [DOI] [PubMed] [Google Scholar]
- Nelson M, Haraldsdottir J. Food photographs: practical guidelines. II. Development and use of photographic atlases for assessing food portion size. Public Health Nutrition. 1998;1(4):231–237. doi: 10.1079/phn19980039. [DOI] [PubMed] [Google Scholar]
- Ngo J, Engelen A, Molag M, Roesle J, Garcia-Segovia P, Serra-Majem L. A review of the use of information and communication technologies for dietary assessment. The British Journal of Nutrition. 2009;101(Suppl. 2):S102–S112. doi: 10.1017/S0007114509990638. [DOI] [PubMed] [Google Scholar]
- Pietinen P, Hartman AM, Haapa E, Räsänen L, Haapakoski J, Palmgren J, Albanes D, Virtamo J, Huttunen JK. Reproducibility and validity of dietary assessment instruments. I. A self-administered food use questionnaire with a portion size picture booklet. American Journal of Epidemiology. 1988a;128(3):655–666. doi: 10.1093/oxfordjournals.aje.a115013. [DOI] [PubMed] [Google Scholar]
- Pietinen P, Hartman AM, Haapa E, Räsänen L, Haapakoski J, Palmgren J, Albanes D, Virtamo J, Huttunen JK. Reproducibility and validity of dietary assessment instruments. II. A qualitative food frequency questionnaire. American Journal of Epidemiology. 1988b;128(3):667–675. doi: 10.1093/oxfordjournals.aje.a115014. [DOI] [PubMed] [Google Scholar]
- Posner BM, Smigelski C, Duggal A, Morgan JL, Cobb J, Cupples LA. Validation of two-dimensional models for estimation of portion size in nutrition research. Journal of the American Dietetic Association. 1992;92(6):738–741. [PubMed] [Google Scholar]
- Thompson FE, Subar AF. Dietary assessment methodology. In: Coulston AM, Rock CL, Monsen ER, editors. Nutrition in the Prevention and Treatment of Disease. Chapter 1. San Diego, CA: Academic Press; 2001. pp. 3–30. [Google Scholar]
- Trabulsi J, Schoeller DA. Evaluation of dietary assessment instruments against doubly labeled water, a biomarker of habitual energy intake. American Journal of Physiology, Endocrinology and Metabolism. 2001;281(5):E891–E899. doi: 10.1152/ajpendo.2001.281.5.E891. [DOI] [PubMed] [Google Scholar]
- Tsai T. A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses. IEEE Journal of Robotics and Automation. 1987;3(4):323–344. [Google Scholar]
- Turconi G, Guarcello M, Gigli Berzolari F, Carolei A, Bazzano R, Roggi C. An evaluation of a color food photography atlas as a tool for quantifying food portion size in epidemiological dietary surveys. European Journal of Clinical Nutrition. 2005;59(8):923–931. doi: 10.1038/sj.ejcn.1602162. [DOI] [PubMed] [Google Scholar]
- Safaee-Rad R, Tchoukanov I, Smith KC, Benhabib B. Three-dimensional location estimation of circular features for machine vision. IEEE Transaction on Robotics Automation. 1992;8(5):624–640. [Google Scholar]
- Shin YC, Ahmad S. 3D location of circular and spherical features by monocular model-based vision. Proceedings of the IEEE International Conference of System, Man and Cybernetics; Boston, USA. 1989. [Google Scholar]
- Sonka M, Hlavac V, Boyle R. Image Processing, Analysis, and Machine Vision. third ed. Toronto, Canada: Thomson Engineering; 2007. [Google Scholar]
- Subar AF, Crafts J, Zimmerman TP, Wilson M, Mittl B, Islam NG, McNutt S, Potischman N, Buday R, Hull SG, Baranowski T, Guenther PM, Willis G, Tapia R, Thompson FE. Assessment of the accuracy of portion size reports using computer-based food photographs aids in the development of an automated self-administered 24-hour recall. Journal of the American Dietetic Association. 2010;110(1):55–64. doi: 10.1016/j.jada.2009.10.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sun M, Fernstrom JD, Jia W, Hackworth SA, Yao N, Li Y, Li C, Fernstrom MH, Sclabassi RJ. A wearable electronic system for objective dietary assessment. Journal of the American Dietetic Association. 2010a;110(1):45–47. doi: 10.1016/j.jada.2009.10.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sun M, Fernstrom JD, Jia W, Yao N, Hackworth SA, Liu X, Li C, Liu Q, Li Y, Fernstrom MH, Sclabassi RJ. Assessment of food intake and physical activity: a computational approach. In: Chen C, editor. Handbook of Pattern Recognition and Computer Vision. fourth ed. Hackensack, NJ, USA: World Scientific Publishing Co.; 2010b. pp. 667–686. [Google Scholar]
- Sun M, Liu Q, Schmidt K, Yang J, Yao N, Fernstrom JD, Fernstrom MH, DeLany JP, Sclabassi RJ. Determination of food portion size by image processing. Proceedings of the IEEE International Conference on Engineering in Medicine and Biology; Vancouver, BC, Canada. 2008. [DOI] [PubMed] [Google Scholar]
- Sun M, Yao N, Hackworth SA, Yang J, Fernstrom JD, Fernstrom MH, Sclabassi RJ. A human-centric smart system assisting people in healthy diet and active living. Proceedings of the International Symposium of Digital Life Technologies: Human-Centric Smart Living Technology; Tainan, Taiwan, China. 2009. [Google Scholar]
- USDA. USDA Food and Nutrient Database for Dietary Studies, 4.1. Beltsville, MD: Agricultural Research Service, Food Surveys Research Group; 2010. [Google Scholar]
- Wacholder S. When measurement errors correlate with truth: surprising effects of nondifferential misclassification. Epidemiology. 1995;6(2):157–161. doi: 10.1097/00001648-199503000-00012. [DOI] [PubMed] [Google Scholar]
- Wang TY, Nguang SK. Low cost sensor for volume and surface area computation of axi-symmetric agricultural products. Journal of Food Engineering. 2007;79(3):870–877. [Google Scholar]
- Williamson DA, Allen R, Martin PD, Alfonso AJ, Gerald B, Hunt A. Comparison of digital photography to weighed and visual estimation of portion sizes. Journal of the American Dietetic Association. 2003;103(9):1139–1145. doi: 10.1016/s0002-8223(03)00974-x. [DOI] [PubMed] [Google Scholar]
- Yue Y, Jia W, Fernstrom JD, Sclabassi RJ, Fernstrom MH, Yao N, Sun M. Food volume estimation using a circular reference in image-based dietary studies. Proceedings of the 35th Northeast Biomedical Engineering Conference; New York, NY, USA. 2010. [Google Scholar]
- Zhu F, Bosch M, Woo I, Kim S, Boushey CJ, Ebert DS, Delp EJ. The use of mobile devices in aiding dietary assessment and evaluation. IEEE Journal of Selected Topics in Signal Processing. 2010;4(4):756–766. doi: 10.1109/JSTSP.2010.2051471. [DOI] [PMC free article] [PubMed] [Google Scholar]