Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Aug 17;16(8):e0255979. doi: 10.1371/journal.pone.0255979

Differential privacy for eye tracking with temporal correlations

Efe Bozkir 1,*,#, Onur Günlü 2,#, Wolfgang Fuhl 1, Rafael F Schaefer 2, Enkelejda Kasneci 1
Editor: Luca Citi3
PMCID: PMC8370645  PMID: 34403454

Abstract

New generation head-mounted displays, such as VR and AR glasses, are coming into the market with already integrated eye tracking and are expected to enable novel ways of human-computer interaction in numerous applications. However, since eye movement properties contain biometric information, privacy concerns have to be handled properly. Privacy-preservation techniques such as differential privacy mechanisms have recently been applied to eye movement data obtained from such displays. Standard differential privacy mechanisms; however, are vulnerable due to temporal correlations between the eye movement observations. In this work, we propose a novel transform-coding based differential privacy mechanism to further adapt it to the statistics of eye movement feature data and compare various low-complexity methods. We extend the Fourier perturbation algorithm, which is a differential privacy mechanism, and correct a scaling mistake in its proof. Furthermore, we illustrate significant reductions in sample correlations in addition to query sensitivities, which provide the best utility-privacy trade-off in the eye tracking literature. Our results provide significantly high privacy without any essential loss in classification accuracies while hiding personal identifiers.

Introduction

Recent advances in the field of head-mounted displays (HMDs), computer graphics, and eye tracking enable easy access to pervasive eye trackers along with modern HMDs. Soon, the usage of such devices might result in a significant increase in the amount of eye movement data collected from users across different application domains such as gaming, entertainment, or education. A large part of this data is indeed useful for personalized experience and user-adaptive interaction. Especially in virtual and augmented reality (VR/AR), it is possible to derive plenty of sensitive information about users from the eye movement data. In general, it has been shown that eye tracking signals can be employed for activity recognition even in challenging everyday tasks [13], to detect cognitive load [4, 5], mental fatigue [6], and many other user states. Similarly, assessment of situational attention [7], expert-novice analysis in areas such as medicine [8] and sports [9], detection of personality traits [10], and prediction of human intent during robotic hand-eye coordination [11] can also be carried out based on eye movement features. Additionally, eye movements are useful for early detection of anomias [12] and diseases [13]. More importantly, eye movement data allow biometric authentication, which is considered to be a highly sensitive task [14]. A task-independent authentication using eye movement features and Gaussian mixtures is, for example, discussed by Kinnunen et al. [15]. Additionally, biometric identification based on an eye movements and oculomotor plant model are introduced by Komogortsev and Holland [16] and by Komogortsev et al. [17]. Eberz et al. [18] discuss that eye movement features can be used reliably also for authentication both in consumer level devices and various real world tasks, whereas Zhang et al. [19] show that continuous authentication using eye movements is possible in VR headsets. While authentication via eye movements could be useful in biometric applications, the applications that do not require any authentication step possess privacy risks for the individuals if such information is not hidden in the data. In addition, if such information is linked to personal identifiers, the risk might be even higher.

As biometric content can be retrieved from eye movements, it is important to protect them against adversarial behaviors such as membership inference. According to Steil et al. [20, p. 3], people agree to share their eye tracking data if a governmental health agency is involved in owning data or if the purpose is research. Therefore, privacy-preserving techniques are needed especially on the data sharing side of eye tracking considering that the usage of VR/AR devices with integrated eye trackers increases. As removing only the personal identifiers from data is not enough for anonymization due to linkage attacks [21], more sophisticated techniques for achieving user level privacy are necessary. Differential privacy [22, 23] is one effective solution, especially in the area of database applications. It protects user privacy by adding randomly generated noise for a given sensitivity and desired privacy parameter, ϵ. The differentially private mechanisms provide aggregate statistics or query answers while protecting the information of whether an individual’s data was contained in a dataset. However, high dimensionality of the data and temporal correlations can reduce utility and privacy, respectively. Since eye movement features are high dimensional, temporally correlated, and usually contain recordings with long durations, it is important to tackle utility and privacy problems simultaneously. For eye movement data collected from HMDs or smart glasses, both local and global differential privacy can be applied. Applying differential privacy mechanisms to eye movement data would optimally anonymize the query outcomes that are carried out on such data while keeping data utility and usability high enough. As opposed to global differential privacy, local differential privacy adds user level noise to the data but assumes that the user sends data to a central data collector after adding local noise [24, 25]. While both could be useful depending on the application use-case, for this work, we focus on global differential privacy, considering that in many VR/AR applications which collect eye movement data, there is a central trusted user-level data collector and publisher.

To apply differential privacy to the eye movement data, we evaluate the standard Laplace Perturbation Algorithm (LPA) [22] of differential privacy and Fourier Perturbation Algorithm (FPA) [26]. The latter is suitable for time series data such as the eye movement feature signals. We propose two different methods that apply the FPA to chunks of data using original eye movement feature signals or consecutive difference signals. While preserving differential privacy using parallel compositions, chunk-based methods decrease query sensitivity and computational complexity. The difference-based method significantly decreases the temporal correlations between the eye movement features in addition to the decorrelation provided by the FPA that uses the discrete Fourier transform (DFT) as, e.g., in the works of Günlü and İşcan [27] and Günlü et al. [28]. The difference-based method provides a higher level of privacy since consecutive sample differences are observed to be less correlated than original consecutive data. Furthermore, we evaluate our methods using differentially private eye movement features in document type, gender, scene privacy sensitivity classification, and person identification tasks on publicly available eye movement datasets by using similar configurations to previous works by Steil et al. [20, 29]. To generate differentially private eye movement data, we use the complete data instead of applying a subsampling step, used by Steil et al. [20] to reduce the sensitivity and to improve the classification accuracies for document type and privacy sensitivity. In addition, the previous work [20] applies the exponential mechanism for differential privacy on the eye movement feature data. The exponential mechanism is useful for situations where the best enumerated response needs to be chosen [30]. In eye movements, we are not interested in the “best” response but in the feature vector. Therefore, we apply the Laplace mechanism. In summary, we are the first to propose differential privacy solutions for aggregated eye movement feature signals by taking the temporal correlations into account, which can help provide user privacy especially for HMD or smart glass usage in VR/AR setups.

Our main contributions are as follows. (1) We propose chunk-based and difference-based differential privacy methods for eye movement feature signals to reduce query sensitivities, computational complexity, and temporal correlations. Furthermore, (2) we evaluate our methods on two publicly available eye movement datasets, i.e., MPIIDPEye [20] and MPIIPrivacEye [29], by comparing them with standard techniques such as LPA and FPA using the multiplicative inverse of the absolute normalized mean square error (NMSE) as the utility metric. In addition, we evaluate document type and gender classification, and privacy sensitivity classification accuracies as classification metrics using differentially private eye movements in the MPIIDPEye and MPIIPrivacEye datasets, respectively. Classification accuracy is used in the literature as a practical utility metric that shows how useful the data and proposed methods are. Our utility metric also provides insights into the divergence trend of differentially private outcomes and is analytically trackable unlike the classification accuracy. For both datasets, we also evaluate person identification task using differentially private data. Our results show significantly better performance as compared to previous works while handling correlated data and decreasing query sensitivities by dividing the data into smaller chunks. In addition, our methods hide personal identifiers significantly better than existing methods.

Previous research

There are few works that focus on privacy-preserving eye tracking. Liebling and Preibusch [31] provide motivation as to why privacy considerations are needed for eye tracking data by focusing on gaze and pupillometry. Practical solutions are; therefore, introduced to protect user identity and sensitive stimuli based on a degraded iris authentication through optical defocus [32] and an automated disabling mechanism for the eye tracker’s ego perspective camera with the help of a mechanical shutter depending on the detection of privacy sensitive content [29]. Furthermore, a function-specific privacy model for privacy-preserving gaze estimation task and privacy-preserving eye videos by replacing the iris textures are proposed by Bozkir and Ünal et al. [33] and by Chaudhary and Pelz [34], respectively. In addition, solutions for privacy-preserving eye tracking data streaming [35] and real-time privacy control for eye tracking systems using area-of-interests [36] are also introduced in the literature. These works lack studying effects of temporal correlations.

For the user identity protection on aggregated eye movement features, works that focus on differential privacy are more relevant for us. Recently, standard differential privacy mechanisms are applied to heatmaps [37] and eye movement data that are obtained from a VR setup [20]. These works do not address the effects of temporal correlations in eye movements over time in the privacy context. In the privacy literature, there are privacy frameworks such as the Pufferfish [38] or the Olympus [39] for correlated and sensor data, respectively. These works, however, have different assumptions. For instance, the Pufferfish requires a domain expert to specify potential secrets and discriminative pairs, and Olympus models privacy and utility requirements as adversarial networks. As our focus is to protect user identity in the eye movements, we opt for differential privacy by discussing the effects of temporal correlations in eye movements over time and propose methods to reduce them. It has been shown that standard differential privacy mechanisms are vulnerable to temporal correlations as such mechanisms assume that data at different time points are independent from each other or adversaries lack the information about temporal correlations, leading an increased privacy loss of a differential privacy mechanism over time due to the temporal correlations [40, 41]. The aggregated eye movement features over time might end up in an extreme case of such correlations due to various user behaviors. Therefore, in addition to discussing the effects of such correlations on differential privacy over time, we propose methods to reduce the correlations so that the privacy leakage due to the temporal correlations are minimal.

Materials and methods

In this section, the theoretical background of differential privacy mechanisms, proposed methods, and evaluated datasets are discussed.

Theoretical background

Differential privacy uses a metric to measure the privacy risk for an individual participating in a database. Considering a dataset with weights of N people and a mean function, when an adversary queries the mean function for N people, the average weight over N people is obtained. After the first query, an additional query for N − 1 people automatically leaks the weight of the remaining person. Using differential privacy, noise is added to the outcome of a function so that the outcome does not significantly change based on whether a randomly chosen individual participated in the dataset. The amount of noise added should be calibrated carefully since a high amount of noise might decrease the utility. We next define differential privacy.

Definition 1. ϵ-Differential Privacy (ϵ-DP) [22, 23]. A randomized mechanism M is ϵ-differentially private if for all databases D and D′ that differ at most in one element for all S ⊆ Range(M) with

Pr[M(D)S]eϵPr[M(D)S]. (1)

The variance of the added noise depends on the query sensitivity, which is defined as follows.

Definition 2. Query sensitivity [22]. For a random query Xn and w ∈ {1, 2}, the query sensitivity Δw of Xn is the smallest number for all databases D and D′ that differ at most in one element such that

||Xn(D)-Xn(D)||wΔw(Xn) (2)

where the Lw-distance is defined as

||Xn||w=i=1n(|Xi|)ww. (3)

We list theorems that are used in the proposed methods.

Theorem 1. Sequential composition theorem [42]. Consider n mechanisms Mi that randomization of each query is independent for i = 1, 2, …, n. If M1, M2, …, Mn are ϵ1, ϵ2, …, ϵn-differentially private, respectively, then their joint mechanism is (i=1ni)-differentially private.

Theorem 2. Parallel composition theorem [42]. Consider n mechanisms as Mi for i = 1, 2, …, n that are applied to disjoint subsets of an input domain. If M1, M2, …, Mn are ϵ1, ϵ2, …, ϵn-differentially private, respectively, then their joint mechanism is maxi1,ni-differentially private.

We define the Laplace Perturbation Algorithm (LPA) [22]. To guarantee differential privacy, the LPA generates the noise according to a Laplace distribution. Lap(λ) denotes a random variable drawn from a Laplace distribution with a probability density function (PDF): Pr[Lap(λ)=h]=12λe-h/λ, where Lap(λ) has zero mean and variance 2λ2. We denote the noisy and differentially private values as X˜i=Xi(D)+Lap(λ) for i = 1, 2, …, n. Since we have a series of eye movement observations, the final noisy eye movement observations are generated as X˜n=Xn(D)+Lapn(λ), where Lapn(λ) is a vector of n independent Lap(λ) random variables and Xn(D) is the eye movement observations without noise. The LPA is ϵ-differentially private for λ = Δ1(Xn)/ϵ [22].

We define the error function that we use to measure the differences between original Xn and noisy X˜n observations. For this purpose, we use the metric normalized mean square error (NMSE) defined as

NMSE=1ni=1n(Xi-X˜i)2X¯X˜¯ (4)

where

X¯=1ni=1nXi,X˜¯=1ni=1nX˜i. (5)

We define the utility metric as

Utility=1|NMSE|. (6)

As differential privacy is achieved by adding random noise to the data, there is a utility-privacy trade-off. Too much noise leads to high privacy; however, it might also result in poor analyses on the further tasks on eye movements. Therefore, it is important to find a good trade-off.

Methods

Standard differential privacy mechanisms are vulnerable to temporal correlations, since the independent noise realizations that are added to temporally correlated data could be useful for adversaries. However, decorrelating the data without the domain knowledge before adding the noise might remove important eye movement patterns and provide poor results in analyses. Many eye movement features are extracted by using time windows, as in previous work [20, 29], which makes the features highly correlated. Another challenge is that the duration of eye tracking recordings could change depending on the personal behaviors, skills, or personalities of the users. The longer duration causes an increased query sensitivity, which means that higher amounts of noise should be added to achieve differential privacy. In addition, when correlations between different data points exist, ϵ′ is defined as the actual privacy metric instead of ϵ [43] that is obtained considering the fact that correlations can be used by an attacker to obtain more information about the differentially private data by filtering. In this work, we discuss and propose generic low-complexity methods to keep ϵ′ small for eye movement feature signals. To deal with correlated eye movement feature signals, we propose three different methods: FPA, chunk-based FPA (CFPA) for original feature signals, and chunk-based FPA for difference based sequences (DCFPA). The sensitivity of each eye movement feature signal is calculated by using the Lw-distance such that

Δwf(Xn)=maxp,qXn,(p,f)-Xn,(q,f)w=maxp,qt=1n(|Xt(p,f)-Xt(q,f)|)ww (7)

where Xn, (p, f) and Xn, (q, f) denote observation vectors for a feature f from two participants p and q, n denotes the maximum length of the observation vectors, and w ∈ {1, 2}.

Fourier Perturbation Algorithm (FPA)

In the FPA [26], the signal is represented with a small number of transform coefficients such that the query sensitivity of the representative signal decreases. A smaller query sensitivity decreases the noise power required to make the noisy signal differentially private. In the FPA, the signal is transformed into the frequency domain by applying Discrete Fourier Transform (DFT), which is commonly applied as a non-unitary transform. The frequency domain representation of a signal consists of less correlated transform coefficients as compared to the time domain signal due to the high decorrelation efficiency of the DFT. Therefore, the correlation between the eye movement feature signals is reduced by applying the DFT. After the DFT, the noise sampled from the LPA is added to the first k elements of DFT(Xn) that correspond to k lowest frequency components, denoted as Fk = DFTk(Xn). Once the noise is added, the remaining part (of size nk) of the noisy signal F˜k is zero padded and denoted as PADn(F˜k). Lastly, using the Inverse DFT (IDFT), the padded signal is transformed back into the time domain. We can show that ϵ-differential privacy is satisfied by the FPA for λ=nkΔ2(Xn) unlike the value claimed in previous work [26], as observed independently by Kellaris and Papadopoulos [44]. The procedure is summarized in Fig 1, and the proof is provided below. Since not all coefficients are used, in addition to the perturbation error caused by the added noise, a reconstruction error caused by the lossy compression is introduced. It is important to determine the number of used coefficients k to minimize the total error. We discuss how we choose k values for FPA-based methods below.

Fig 1. Flow of the Fourier Perturbation Algorithm (FPA).

Fig 1

Proof of FPA being differentially private. We next prove that the FPA is ϵ differentially private for λ=(nkΔ2(Xn))/. First, we prove the inequalities (a) and (b) in the following.

Δ1(F^n)(a)k·Δ2(F^n)(b)n·k·Δ2(Xn) (8)

where F^n(I)=[F^k(I),0,0,,0] such that nk zeros are padded. Consider (8)(a), which follows since we have

Δ1(F^n)=maxI,IF^n(I)-F^n(I)1=maxI,Ij=1n|F^j(I)-F^j(I)|=maxI,Ij=1k|F^j(I)-F^j(I)|·1 (9)

so that by applying Cauchy-Schwarz inequality, we obtain

maxI,Ij=1k|F^j(I)-F^j(I)|·1maxI,I(j=1k|F^j(I)-F^j(I)|2)1/2·(j=1k12)1/2maxI,IF^n(I)-F^n(I)2·kk·Δ2(F^n). (10)

Consider next (8)(b), which follows since we obtain

Δ2(F^n)=maxI,IF^n(I)-F^n(I)2=maxI,I(j=1n|F^j(I)-F^j(I)|2)1/2 (11)

and since Fn has more non-zero elements than F^n, we have

Δ2(F^n)maxI,I(j=1n|Fj(I)-Fj(I)|2)1/2. (12)

Recall that Fn(I) = DFT(Xn(I)), Fn(I′) = DFT(Xn(I′)), and DFT is linear, so we have

DFT(Xn(I)-Xn(I))=Fn(I)-Fn(I). (13)

By applying Parseval’s theorem to the DFT, we obtain

(1n·j=1n|Fj(I)-Fj(I)|2)1/2=(j=1n|Xj(I)-Xj(I)|2)1/2. (14)

Combining (12) and (14), we prove (8)(b) since we have

Δ2(F^n)maxI,Ij=1n|Xj(I)-Xj(I)|2·nmaxI,IXn(I)-Xn(I)2·nΔ2(Xn)·n. (15)

Finally, since the LPA that is applied to F^k is ϵ-DP for λ=Δ1(F^n) [22], (8) proves that the FPA is ϵ-DP for λ=nkΔ2(Xn).

Chunk-based FPA (CFPA)

One drawback of directly applying the FPA to the eye movement feature signals is large query sensitivities for each feature f due to long signal sizes. To solve this, Steil et al. [20] propose to subsample the signal using non-overlapping windows, which means removing many data points. While subsampling decreases the query sensitivities, it also decreases the amount of data. Instead, we propose to split each signal into smaller chunks and apply the FPA to each chunk so that complete data can be used. We choose the chunk sizes of 32, 64, and 128 since there are divide-and-conquer type tree-based implementation algorithms for fast DFT calculations when the transform size is a power of 2 [45]. When the signals are split into chunks, chunk level query sensitivities are calculated and used rather than the sensitivity of the whole sequence. Differential privacy for the complete signal is preserved by Theorem 2 [42] since the used chunks are non-overlapping. As the chunk size decreases, the chunk level sensitivity decreases as well as the computational complexity. However, the parameter ϵ′ that accounts for the sample correlations might increase with smaller chunk sizes because temporal correlations between neighboring samples are larger in an eye movement dataset. On the other hand, if the chunk sizes are kept large, then the required amount of noise to achieve differential privacy increases due to the increased query sensitivity. Therefore, a good trade-off between computational complexity, and correlations is needed to determine the optimal chunk size.

Difference- and chunk-based FPA (DCFPA)

To tackle temporal correlations, we convert the eye movement feature signals into difference signals where differences between consecutive eye movement features are calculated as

X^t(f)={Xt(f)-Xt-1(f)}|t=2n,X^1(f)=X1(f). (16)

Using the difference signals denoted by X^n,(f), we aim to further decrease the correlations before applying a differential privacy method. We conjecture that the ratio ϵ′/ϵ decreases in the difference-based method as compared to the FPA method. To support this conjecture, we show that the correlations in the difference signals decrease significantly as compared to the original signals. This results in lower ϵ′ and better privacy for the same ϵ. The difference-based method is applied together with the CFPA. Therefore, the differences are calculated inside chunks. The first element of each chunk is preserved. Then, the FPA mechanism is applied to the difference signals by using query sensitivities calculated based on differences and chunks. For each chunk, noisy difference observations are aggregated to obtain the final noisy signals. This mechanism is differentially private by Theorem 1 [42], and described in Algorithm 1.

Algorithm 1: DCFPA.

Input:Xn, λ, k

Outut: X˜n

1) X^t={Xt-Xt-1}|t=2n,X^1=X1.

2) X^˜n=FPA(X^n,λ,k).

3) Xt˜={X^t˜+X^t-1˜}|t=2n,X˜1=X^1˜.

Since Theorem 1 can be applied to the DCFPA when the consecutive differences are assumed to be independent, which is a valid assumption for eye movement feature signals as we illustrate below, there is also a trade-off between the chunk sizes and utility for the DCFPA. If a large chunk size is chosen, then the total ϵ value could be very large, which reduces privacy. Therefore, we choose chunk sizes of 32, 64, and 128 for the DCFPA as well for evaluation We illustrate the CFPA and DCFPA in Fig 2, for instance with three chunks.

Fig 2. Flow of the CFPA and DCFPA.

Fig 2

Choice of the number of transform coefficients

The proposed methods require a selection of a value for k. A small k value increases the reconstruction error, while a large k value results in an increase in the perturbation error. Therefore, it is important to find an optimal k value that minimizes the sum of the two errors. In this work, we compare a large set of possible k values to choose the best values.

We apply the aforementioned differential privacy mechanisms by using 100 noisy evaluations to find optimal k values applied to features or chunks. Optimal k values have the minimum absolute NMSE for each chunk, eye movement feature, and document or recording type. In a distributed setting, each party should know the k values in advance. However, in a centralized setting, it is crucial to choose the k values in a differentially private manner. To evaluate the differential privacy in the eye tracking area while taking the temporal correlations into account, we focus on optimal k values for this work. One shortcoming of this approach is that the optimal k value compromises some information about the data, which leaks privacy [26]. Our observation is that the information leaked by optimizing the parameter k is negligible as compared to the privacy reduction due to temporally correlated data. Thus, we illustrate the results with optimal k values.

Datasets

We evaluate our methods on two different publicly available eye movement datasets namely, MPIIDPEye and MPIIPrivacEye that are dedicated to privacy-preserving eye tracking. Both datasets consist of aggregated and timely eye movement feature signals related to eye fixations, saccades, blinks, and pupil diameters which are commonly used in VR/AR applications as they represent individual user behaviors. As all minimum values of wordbook features ranging from 1 to 4 are zeros in both datasets, we exclude them from the utility and privacy calculations. In addition, we remark that both datasets are available for non-commercial scientific purposes.

MPIIDPEye [20]: A publicly available eye movement dataset consisting of 60 recordings that is collected from VR devices for a reading task of three document types (comics, newspaper, and textbook) from 20 (10 female, 10 male) participants. Each recording consists of 52 eye movement feature sequences computed with a sliding window size of 30 seconds and a step size of 0.5 seconds.

MPIIPrivacEye [29]: A publicly available eye movement dataset consisting of 51 recordings from 17 participants for 3 consecutive sessions with a head-mounted eye tracker and a field camera, which is similar to an AR setup. Each recording consists of 52 eye movement feature sequences computed with a sliding window size of 30 seconds and a step size of 1 second, and each observation is annotated with binary privacy sensitivity levels of the scene that is being viewed. The dataset also consists of scene features extracted with convolutional neural networks. We do not evaluate the last part of the recording 1 of the participant 10, as the eye movement features are not available for this region. To detect the privacy level of the scene that is being viewed, we remark that the scene is very important [46]; however, an individual’s eye movements can improve the detection rate when they are fused with the information from the scene.

Results

This section discusses data correlations in addition to evaluations using utility and classification metrics. The utility and classification results are averaged over 100 noisy evaluations with the optimal k values in MATLAB. We evaluate and compare the utility of differentially private eye movement feature signals by using absolute NMSE, as this metric provides analytically trackable results. However, it does not provide implications regarding the practical usability of the private eye movement signals. Therefore, we also report classification accuracies of document type, scene privacy sensitivity, gender prediction, and person identification tasks in order to show the usability of the private data and proposed methods. An optimal trade-off between utility tasks (e.g., low absolute NMSE, high classification accuracy in document type prediction) and privacy (e.g., low ϵ, low classification accuracy in person identification or gender prediction tasks) is favorable.

Correlation analysis

Using the correlation coefficient as the metric, we first illustrate high temporal correlation between eye movement feature data. Since there are 52 eye movement features in both datasets, it is not feasible to illustrate all correlation results. Thus, in the following we illustrate the correlations for the features ratio large saccade and blink rate in the MPIIDPEye and MPIIPrivacEye datasets, respectively. The correlation coefficients of ratio large saccade and blink rate for three document and recording types over a time difference Δt w.r.t. the signal samples at, e.g., the fifth time instance for original eye movement feature signals and difference signals for all participants for both datasets are depicted in Figs 36, respectively. As correlations between the difference signals are significantly smaller than correlations between the original eye movement feature signals, the DCFPA is less vulnerable to privacy reduction due to temporal correlations, thus ensuring that the value of ϵ′ is close to the differential privacy design parameter ϵ.

Fig 3. Correlation coefficients of the raw signals of feature ratio large saccade in the MPIIDPEye dataset.

Fig 3

The values are calculated over a time difference of Δt (Each time step corresponds to 0.5s) w.r.t. the samples at the fifth time instance.

Fig 6. Correlation coefficients of the difference signals of feature blink rate in the MPIIPrivacEye dataset.

Fig 6

The values are calculated over a time difference of Δt (Each time step corresponds to 1s) w.r.t. the samples at the fifth time instance.

Utility results

We evaluate the utility defined in Eq (6) by applying our methods separately to different document and recording types; therefore, we report the utility results separately. As we apply the proposed methods separately to each eye movement feature, we first calculate the mean utility of each feature and then calculate the average utility over all features. The utility results for various ϵ values for aforementioned methods on the MPIIDPEye and MPIIPrivacEye datasets are given in Figs 712, respectively.

Fig 7. Utility of the LPA and FPA for MPIIDPEye.

Fig 7

Fig 12. Utility of the DCFPA for MPIIPrivacEye.

Fig 12

Fig 8. Utility of the CFPA for MPIIDPEye.

Fig 8

Fig 9. Utility of the DCFPA for MPIIDPEye.

Fig 9

Fig 10. Utility of the LPA and FPA for MPIIPrivacEye.

Fig 10

Fig 11. Utility of the CFPA for MPIIPrivacEye.

Fig 11

While a high absolute NMSE, i.e., low utility, does not necessarily mean that a mechanism is completely useless, higher utility means that the mechanism would perform more effectively than low utility in various tasks. The trend in the utility results of both evaluated datasets are similar. As the query sensitivities are lower in CFPA, utilities of CFPA are always higher than the utilities of the FPA as theoretically expected. The DCFPA particularly with small chunks outperforms other methods in the most private settings, namely in the lowest ϵ regions. When different chunk sizes are compared within the CFPA and DCFPA, different chunk sizes perform similarly for the CFPA method. For the DCFPA, there is a significant trend for better utilities when the chunk sizes are decreased. However, as temporal correlations in the smaller chunk sizes higher and since a higher chunk size reduces the temporal correlations better, it is ideal to use a higher chunk size if the utilities are comparable. In general, while the LPA, namely the standard Laplace mechanism used for differential privacy, is vulnerable to temporal correlations [41], our methods also outperform it in terms of utilities. In addition to high utilities, the calculation complexities are decreased with the CFPA and DCFPA which is another advantage of chunk-based methods.

Classification accuracy results

We evaluate document type and gender classification results for the MPIIDPEye and privacy sensitivity classification results for the MPIIPrivacEye by using differentially private data generated by the methods which handle temporal correlations in the privacy context. In addition, for both datasets, we evaluate person identification tasks. While a NMSE-based utility metric provides analytically trackable way for comparison, evaluating private data using classification accuracies give insights about the usability of the noisy data in practice. Instead of only using Support Vector Machines (SVM) as in previous works [20, 29], we evaluate a set of classifiers including SVMs, decision trees (DTs), random forests (RFs), and k-Nearest Neighbors (k-NNs). We employ a similar setup as in previous work [20] with radial basis function (RBF) kernel, bias parameter of C = 1, and automatic kernel scale for the SVMs. For RFs and k-NNs, we use 10 trees and k = 11 with a random tie breaker among tied groups, respectively. We normalize the training data to zero mean and unit variance, and apply the same parameters to the test data. Although we do not apply subsampling while generating the differentially private data, which is applied in previous work [20], we use subsampled data for training and testing for document type, gender, and privacy sensitivity classification tasks with window sizes of 10 and 20 for MPIIDPEye and MPIIPrivacEye, respectively, to have a fair comparison and similar amount of data. Apart from the person identification task, all the classifiers are trained and tested in a leave-one-person-out cross-validation setup, which is considered as a more challenging but generic setup. For the person identification task, since it is not possible to carry out the experiments in a leave-one-person-out cross-validation setup, we opt for a similar configuration as in previous work [20] by using the first halves of the signals as training data and the remaining parts as test data. Such setup can be considered as one of the hypothetical best-case scenarios for an adversary as this simulates some set of prior knowledge for an adversary on participants’ visual behaviors. For the person identification task, in order to use similar amount of data with other classification tasks from each signal, we use window sizes of 5 and 10 for MPIIDPEye and MPIIPrivacEye, respectively. For the MPIIDPEye, we evaluate results both with majority voting by summarizing classifications from different time instances for each participant and recording and without majority voting. Privacy sensitivity classification tasks for MPIIPrivacEye are carried only without majority voting since privacy sensitivity of the scene can change at each time step and applying majority voting to such task in our setup is not reasonable.

While classification results cannot be treated directly as the utility, they provide insights into the usability of the differentially private data in practice. We first evaluate document type classification task in the majority voting setting in Table 1 for MPIIDPEye dataset as it is possible to compare our results with the previous work [20]. As previous results quickly drop to the 0.33 guessing probability in high privacy regions, we significantly outperform them particularly with DCFPA and FPA with the accuracies over 0.60 and 0.85, respectively. In the less private regions towards ϵ = 48, this trend still exists with the CFPA and FPA with accuracy results over 0.7 and 0.85. Chunk-based methods perform slightly worse than the FPA in the document type classifications even though the utility of the FPA is lower. We observe that the reading patterns are hidden easier with chunk-based methods; therefore, document type classification task becomes more challenging. This is especially validated with DCFPA methods using different chunk sizes, as DCFPA-128 outperforms smaller chunk-sized DCFPAs even though the sensitivities are higher. Therefore, we conclude that the differential privacy method should be selected for eye movements depending on the further task which will be applied. The document type classification results without majority voting are provided in the table in S1 Table.

Table 1. Document type classification accuracies in the MPIIDPEye dataset using differentially private eye movement features with majority voting.

Document type classification accuracies (k-NN|SVM|DT|RF)
Method ϵ = 0.48 ϵ = 2.4 ϵ = 4.8 ϵ = 24 ϵ = 48
FPA 0.50|0.63|0.82|0.87 0.51|0.63|0.81|0.87 0.5|0.61|0.81|0.87 0.52|0.63|0.82|0.87 0.52|0.64|0.83|0.88
CFPA-32 0.39|0.37|0.45|0.44 0.40|0.38|0.45|0.44 0.40|0.44|0.46|0.44 0.58|0.58|0.55|0.60 0.71|0.69|0.66|0.66
CFPA-64 0.41|0.36|0.45|0.45 0.40|0.37|0.44|0.45 0.40|0.41|0.44|0.45 0.57|0.59|0.55|0.59 0.70|0.70|0.66|0.66
CFPA-128 0.36|0.33|0.45|0.45 0.36|0.33|0.44|0.44 0.37|0.35|0.44|0.45 0.52|0.56|0.52|0.57 0.69|0.68|0.64|0.66
DCFPA-32 0.51|0.37|0.46|0.44 0.51|0.36|0.47|0.42 0.47|0.35|0.47|0.43 0.49|0.37|0.46|0.44 0.48|0.36|0.47|0.45
DCFPA-64 0.61|0.45|0.43|0.41 0.55|0.35|0.43|0.41 0.56|0.41|0.43|0.41 0.60|0.43|0.45|0.42 0.59|0.40|0.44|0.43
DCFPA-128 0.64|0.45|0.46|0.48 0.62|0.42|0.45|0.46 0.69|0.50|0.44|0.46 0.57|0.45|0.45|0.46 0.60|0.42|0.45|0.46

Next, we analyze the gender classification results for MPIIDPEye. All methods are able to hide the gender information in the high privacy regions as it is already challenging to identify it with clean data as accuracies are ≈0.7 in previous work [20]. While we obtain similar results compared to previous work for the gender classification task, the CFPA method is able to predict gender information correctly in the less private regions, namely ϵ = 48, as it also has the highest utility values in these regions. The FPA applied to the complete signal and the DCFPA are not able to classify genders accurately. We observe that higher amount of noise that is needed by the FPA and noising the fine-grained “difference” information between eye movement observations with DCFPA are the reasons for hiding the gender information successfully in all privacy regions. Overall, the CFPA provides an optimal equilibrium between gender and document type classification success in the less private regions if gender information is not considered as a feature that should be protected from adversaries. Otherwise, all proposed methods are able to hide gender information from the data in the higher privacy regions as expected. Gender classification results are depicted in Table 2. Especially in some methods with k-NNs and SVMs, gender classification accuracies are close to zero because of the majority voting and it is validated by the results without majority voting in the table in S2 Table.

Table 2. Gender classification accuracies in the MPIIDPEye dataset using differentially private eye movement features with majority voting.

Gender classification accuracies (k-NN|SVM|DT|RF)
Method ϵ = 0.48 ϵ = 2.4 ϵ = 4.8 ϵ = 24 ϵ = 48
FPA 0.44|0.30|0.43|0.38 0.45|0.30|0.41|0.37 0.44|0.28|0.41|0.39 0.43|0.27|0.43|0.38 0.44|0.31|0.42|0.39
CFPA-32 0.04|0.01|0.26|0.24 0.05|0.01|0.27|0.25 0.05|0.02|0.28|0.27 0.36|0.30|0.50|0.45 0.62|0.50|0.67|0.53
CFPA-64 0.08|0.05|0.27|0.26 0.08|0.04|0.28|0.27 0.10|0.06|0.31|0.27 0.38|0.34|0.52|0.47 0.62|0.51|0.68|0.54
CFPA-128 0.18|0.15|0.32|0.30 0.16|0.12|0.31|0.30 0.18|0.10|0.32|0.31 0.36|0.30|0.50|0.46 0.60|0.47|0.68|0.54
DCFPA-32 0.03|≈0|0.22|0.31 0.04|≈0|0.23|0.32 0.04|≈0|0.22|0.32 0.04|≈0|0.23|0.31 0.04|≈0|0.23|0.32
DCFPA-64 0.04|≈0|0.30|0.33 0.04|≈0|0.30|0.34 0.04|≈0|0.30|0.32 0.04|≈0|0.29|0.34 0.03|≈0|0.30|0.34
DCFPA-128 0.09|0.01|0.34|0.35 0.08|≈0|0.32|0.34 0.08|0.01|0.32|0.35 0.07|≈0|0.33|0.34 0.07|0.01|0.34|0.34

Lastly for the MPIIDPEye, we evaluate person identification task using differentially private data. The resulting classification accuracies with majority voting are depicted in Table 3. By using the FPA, it is possible to identify the participants very accurately, which means that even though the document type classification accuracies of the FPA are higher than the others, a strong adversary can also identify personal ids when this method is used. The same trend also exists in the without majority voting setting, which is reported in the table in S3 Table. The CFPA and DCFPA perform well against person identification attempts in the high privacy regions. However, when the CFPA is used, it is possible to identify personal ids in the less private regions. Overall, for the MPIIDPEye dataset, the DCFPA performs better than the others due to its resistance against person identification and gender classification and relatively high classification accuracies for the document type predictions. We conclude that this is due to the robust decorrelation effect of the DCFPA.

Table 3. Person identification accuracies in the MPIIDPEye dataset using differentially private eye movement features with majority voting.

Person identification accuracies (k-NN|SVM|DT|RF)
Method ϵ = 0.48 ϵ = 2.4 ϵ = 4.8 ϵ = 24 ϵ = 48
FPA 1 1 1 1 1
CFPA-32 0.15|0.08|0.44|0.37 0.13|0.08|0.46|0.39 0.11|0.08|0.48|0.41 0.40|0.11|0.64|0.70 0.72|0.16|0.87|0.92
CFPA-64 0.14|0.08|0.42|0.34 0.13|0.08|0.44|0.37 0.12|0.08|0.45|0.38 0.39|0.11|0.63|0.71 0.70|0.17|0.85|0.92
CFPA-128 0.16|0.05|0.39|0.36 0.15|0.05|0.41|0.36 0.17|0.05|0.43|0.39 0.45|0.07|0.55|0.63 0.70|0.16|0.74|0.88
DCFPA-32 0.06|0.10|0.39|0.37 0.06|0.10|0.39|0.36 0.08|0.10|0.39|0.36 0.10|0.10|0.39|0.37 0.10|0.10|0.40|0.38
DCFPA-64 0.10|0.10|0.33|0.35 0.10|0.10|0.32|0.34 0.10|0.10|0.32|0.33 0.13|0.10|0.31|0.34 0.13|0.10|0.32|0.33
DCFPA-128 0.09|0.05|0.24|0.28 0.09|0.05|0.25|0.27 0.10|0.05|0.23|0.27 0.10|0.06|0.24|0.26 0.10|0.05|0.22|0.25

For the MPIIPrivacEye, we report privacy sensitivity classification accuracies using differentially private eye movement features in the Table 4. The FPA performs worse than our methods. The DCFPA, particularly with the chunk size of 32, outperforms all other methods slightly in the higher privacy regions as it is also the case for the utility results. In the lower privacy regions, the CFPA performs the best with ≈0.60 accuracy. Since performance does not drop significantly in the higher chunk sizes, it is reasonable to use higher chunk-sized methods as they decrease the temporal correlations better. While having an accuracy of approximately 0.6 in a binary classification problem does not form the best performance, according to the previous work [29], privacy sensitivity classification using only eye movements with clean data in a person-independent setup only performs marginally higher than 0.60. Therefore, we show that even though we use differentially private data in the most private settings, we obtain similar results to the classification results using clean data. This means that differentially private eye movements can be used along with scene features for detecting privacy sensitive scenes in AR setups.

Table 4. Privacy sensitivity classification accuracies in the MPIIPrivacEye dataset using differentially private eye movement features.

Privacy sensitivity classification accuracies (k-NN|SVM|DT|RF)
Method ϵ = 0.48 ϵ = 2.4 ϵ = 4.8 ϵ = 24 ϵ = 48
FPA 0.49|0.58|0.51|0.55 0.49|0.58|0.51|0.55 0.49|0.58|0.51|0.55 0.50|0.58|0.51|0.55 0.50|0.59|0.51|0.55
CFPA-32 0.55|0.59|0.52|0.56 0.55|0.58|0.52|0.56 0.55|0.58|0.52|0.56 0.56|0.58|0.53|0.57 0.58|0.60|0.54|0.58
CFPA-64 0.55|0.58|0.52|0.56 0.55|0.58|0.52|0.56 0.55|0.58|0.52|0.56 0.56|0.58|0.53|0.57 0.58|0.59|0.54|0.58
CFPA-128 0.55|0.57|0.52|0.56 0.55|0.57|0.52|0.56 0.55|0.57|0.52|0.56 0.56|0.58|0.53|0.57 0.58|0.59|0.54|0.59
DCFPA-32 0.54|0.59|0.52|0.56 0.55|0.59|0.52|0.56 0.55|0.59|0.52|0.56 0.54|0.59|0.52|0.56 0.55|0.59|0.52|0.56
DCFPA-64 0.54|0.58|0.52|0.56 0.54|0.58|0.52|0.56 0.54|0.58|0.52|0.56 0.54|0.58|0.52|0.56 0.54|0.58|0.52|0.56
DCFPA-128 0.54|0.57|0.52|0.56 0.54|0.57|0.52|0.56 0.54|0.57|0.52|0.56 0.54|0.57|0.52|0.56 0.54|0.57|0.52|0.56

The results of the person identification task in the MPIIPrivacEye dataset are similar to the results of the MPIIDPEye dataset and the results with majority voting are depicted in Table 5. Personal identifiers are predicted very accurately when the FPA is used. The CFPA and DCFPA are resistant to person identification attacks in all privacy regions performing around the random guess probability in almost all cases. Similar to the classification results of the MPIIDPEye dataset, the DCFPA method performs the best when utility-privacy trade-off is taken into consideration. The person identification results without majority voting are presented in the table in S4 Table.

Table 5. Person identification classification accuracies in the MPIIPrivacEye dataset using differentially private eye movement features with majority voting.

Person identification classification accuracies (k-NN|SVM|DT|RF)
Method ϵ = 0.48 ϵ = 2.4 ϵ = 4.8 ϵ = 24 ϵ = 48
FPA 1 1 1 1 1
CFPA-32 0.05|0.06|0.07|0.07 0.05|0.06|0.07|0.07 0.06|0.06|0.08|0.07 0.07|0.06|0.09|0.11 0.11|0.06|0.14|0.16
CFPA-64 0.06|0.06|0.06|0.07 0.06|0.06|0.06|0.06 0.06|0.06|0.07|0.07 0.07|0.06|0.09|0.09 0.11|0.06|0.16|0.16
CFPA-128 0.06|0.06|0.07|0.07 0.06|0.06|0.07|0.07 0.06|0.06|0.07|0.08 0.07|0.06|0.09|0.11 0.11|0.06|0.15|0.15
DCFPA-32 0.06|0.05|0.08|0.07 0.06|0.06|0.07|0.08 0.07|0.05|0.08|0.08 0.07|0.05|0.08|0.08 0.07|0.06|0.08|0.08
DCFPA-64 0.06|0.05|0.06|0.06 0.06|0.05|0.06|0.06 0.06|0.05|0.06|0.06 0.05|0.06|0.06|0.06 0.06|0.05|0.06|0.06
DCFPA-128 0.05|0.05|0.06|0.06 0.05|0.05|0.05|0.06 0.06|0.05|0.06|0.06 0.05|0.05|0.05|0.06 0.06|0.05|0.05|0.06

Discussion

We compared our differential privacy methods with the standard Laplace mechanism as well as the FPA method, which is proposed for temporally correlated data, by using the MPIIDPEye and MPIIPrivacEye datasets. The utility results based on the NMSE metric show that due to the reduced sensitivities as a result of the chunking operations, the CFPA and DCFPA perform better than the FPA and standard Laplace mechanism. While larger chunk sizes applied with the CFPA and DCFPA decrease the effects of temporal correlations on the differential privacy mechanisms, they also increase the sensitivities, leading to higher amount of noise addition to the data and worse utility performance. Utility evaluations represent how much differentially private signals diverge from the original signals. Having eye movement feature signals less diverged from the original values by providing the differential privacy would lead better performance in various tasks. While the FPA, CFPA, and DCFPA are appropriate for temporally correlated data, the DCFPA uses the consecutive differences of eye movement feature signals, which are significantly less correlated than the original feature signals, as illustrated in Figs 4 and 6. Thus, the DCFPA is less vulnerable to temporal correlations in the differential privacy context.

Fig 4. Correlation coefficients of the difference signals of feature ratio large saccade in the MPIIDPEye dataset.

Fig 4

The values are calculated over a time difference of Δt (Each time step corresponds to 0.5s) w.r.t. the samples at the fifth time instance.

In addition to utility results, we evaluated document type, gender, and person identification tasks for the MPIIDPEye dataset and privacy sensitivity classification of the observed scene and person identification task for the MPIIPrivacEye dataset and compared our results with the previous works especially in the eye tracking literature. The FPA outperforms the CFPA and DCFPA in document type classification task since the chunks perturb “Z”-type reading patterns. However, this might be a task-specific outcome as the CFPA and DCFPA perform better in terms of utility. In addition, when the FPA is used, personal identifiers can be estimated with high accuracies in both datasets. On the contrary, especially the DCFPA provides decreased probabilities for person identification tasks in the MPIIDPEye, and probabilities close to the random guess probability for the MPIIPrivacEye, which are optimal from a privacy-preservation perspective. We remark that this outcome is also related to decreased temporal correlations. Gender information is successfully hidden in all methods and scene privacy can be predicted to some extent using differentially private eye movement signals. In addition, privacy sensitivity detection results on the MPIIPrivacEye are consistent with the utility results based on the NMSE metric.

Due to the significant reduction of temporal correlations and high utility and relatively accurate classification results in different tasks, the DCFPA is the best performing differential privacy method for eye movement feature signals. In addition, it is not possible to recognize the person accurately from eye movement data when the DCFPA is used. From correlation reduction point of view, in both methods namely, CFPA and DCFPA, when the performances are similar, it is reasonable to use higher chunk sizes as such chunks are less vulnerable to temporal correlations as illustrated in Figs 3 and 5. Overall, our methods outperform the state-of-the-art for differential privacy for aggregated eye movement feature signals.

Fig 5. Correlation coefficients of the raw signals of feature blink rate in the MPIIPrivacEye dataset.

Fig 5

The values are calculated over a time difference of Δt (Each time step corresponds to 1s) w.r.t. the samples at the fifth time instance.

Conclusion

We proposed different methods to achieve differential privacy for eye movement feature signals by correcting, extending, and adapting the FPA method. Since eye movement features are correlated over time and are high dimensional, standard differential privacy methods provide low utility and are vulnerable to inference attacks. Thus, we proposed privacy solutions for temporally correlated eye movement data. Our methods can be easily applied to other biometric human-computer interaction data as well since they are independent of the used data and outperform the state-of-the-art methods in terms of both NMSE and classification accuracy and reduce the correlations significantly. In future work, we will analyze the actual privacy metric ϵ′ which takes the data correlations into account and choose k values in a private manner for the centralized differential privacy setting.

Supporting information

S1 Table. Document type classification results without majority voting for the MPIIDPEye dataset.

(PDF)

S2 Table. Gender classification results without majority voting for the MPIIDPEye dataset.

(PDF)

S3 Table. Person identification results without majority voting for the MPIIDPEye dataset.

(PDF)

S4 Table. Person identification results without majority voting for MPIIPrivacEye dataset.

(PDF)

Acknowledgments

O. Günlü thanks Ravi Tandon for his useful suggestions. E. Bozkir thanks Martin Pawelczyk and Mete Akgün for useful discussions.

Data Availability

Relevant data files are provided via following url: https://atreus.informatik.uni-tuebingen.de/bozkir/dp_eye_tracking.

Funding Statement

O. Günlü and R. F. Schaefer are supported by the German Federal Ministry of Education and Research (BMBF) within the national initiative for “Post Shannon Communication (NewCom)” under the Grant 16KIS1004. We acknowledge support by Open Access Publishing Fund of University of Tübingen. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Steil J, Bulling A. Discovery of Everyday Human Activities from Long-term Visual Behaviour Using Topic Models. In: ACM International Joint Conference on Pervasive and Ubiquitous Computing. New York, NY, USA: ACM; 2015. p. 75–85.
  • 2.Braunagel C, Geisler D, Rosenstiel W, Kasneci E. Online Recognition of Driver-Activity Based on Visual Scanpath Classification. IEEE Intelligent Transportation Systems Magazine. 2017;9(4):23–36. doi: 10.1109/MITS.2017.2743171 [DOI] [Google Scholar]
  • 3.Ishimaru S, Kunze K, Kise K, Weppner J, Dengel A, Lukowicz P, et al. In the Blink of an Eye: Combining Head Motion and Eye Blink Frequency for Activity Recognition with Google Glass. In: ACM Augmented Human International Conference. New York, NY, USA: ACM; 2014. p. 15:1–15:4.
  • 4.Appel T, Scharinger C, Gerjets P, Kasneci E. Cross-subject Workload Classification Using Pupil-related Measures. In: ACM Symposium on Eye Tracking Research & Applications. New York, NY, USA: ACM; 2018. p. 4:1–4:8.
  • 5.Krejtz K, Duchowski AT, Niedzielska A, Biele C, Krejtz I. Eye Tracking Cognitive Load Using Pupil Diameter and Microsaccades with Fixed Gaze. PLOS ONE. 2018;13(9):1–23. doi: 10.1371/journal.pone.0203629 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Yamada Y, Kobayashi M. Detecting Mental Fatigue from Eye-tracking Data Gathered While Watching Video: Evaluation in Younger and Older Adults. Artificial Intelligence in Medicine. 2018;91:39–48. doi: 10.1016/j.artmed.2018.06.005 [DOI] [PubMed] [Google Scholar]
  • 7.Bozkir E, Geisler D, Kasneci E. Assessment of Driver Attention During a Safety Critical Situation in VR to Generate VR-based Training. In: ACM Symposium on Applied Perception 2019. New York, NY, USA: ACM; 2019. p. 23:1–23:5.
  • 8.Castner N, Kasneci E, Kübler T, Scheiter K, Richter J, Eder T, et al. Scanpath Comparison in Medical Image Reading Skills of Dental Students: Distinguishing Stages of Expertise Development. In: ACM Symposium on Eye Tracking Research & Applications. New York, NY, USA: ACM; 2018. p. 39:1–39:9.
  • 9.van Leeuwen PM, de Groot S, Happee R, de Winter JCF. Differences Between Racing and Non-racing Drivers: A Simulator Study Using Eye-tracking. PLOS ONE. 2017;12(11):1–19. doi: 10.1371/journal.pone.0186871 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Berkovsky S, Taib R, Koprinska I, Wang E, Zeng Y, Li J, et al. Detecting Personality Traits Using Eye-Tracking Data. In: ACM Conference on Human Factors in Computing Systems. New York, NY, USA: ACM; 2019. p. 221:1–221:12.
  • 11.Razin Y, Feigh K. Learning to Predict Intent from Gaze During Robotic Hand-Eye Coordination. In: AAAI Conference on Artificial Intelligence. Palo Alto, CA, USA: AAAI Press; 2017. p. 4596–4602.
  • 12.Ungrady MB, Flurie M, Zuckerman BM, Mirman D, Reilly J. Naming and Knowing Revisited: Eyetracking Correlates of Anomia in Progressive Aphasia. Frontiers in Human Neuroscience. 2019;13:354. doi: 10.3389/fnhum.2019.00354 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Fernández G, Manes F, Politi L, Orozco D, Schumacher M, Castro L, et al. Patients with Mild Alzheimer’s Disease Fail When Using Their Working Memory: Evidence from the Eye Tracking Technique. Journal of Alzheimer’s Disease. 2015;50. doi: 10.3233/JAD-150265 [DOI] [PubMed] [Google Scholar]
  • 14.Günlü O. Key Agreement with Physical Unclonable Functions and Biometric Identifiers [Ph.D. Thesis]. Technical University of Munich. Munich, Germany; 2018. [Google Scholar]
  • 15.Kinnunen T, Sedlak F, Bednarik R. Towards Task-independent Person Authentication Using Eye Movement Signals. In: ACM Symposium on Eye-Tracking Research & Applications. New York, NY, USA: ACM; 2010. p. 187–190.
  • 16.Komogortsev OV, Holland CD. Biometric authentication via complex oculomotor behavior. In: 2013 IEEE Sixth International Conference on Biometrics: Theory, Applications and Systems. New York, NY, USA: IEEE; 2013. p. 1–8.
  • 17.Komogortsev OV, Jayarathna S, Aragon CR, Mahmoud M. Biometric Identification via an Oculomotor Plant Mathematical Model. In: ACM Symposium on Eye-Tracking Research & Applications. New York, NY, USA: ACM; 2010. p. 57–60.
  • 18.Eberz S, Rasmussen KB, Lenders V, Martinovic I. Looks Like Eve: Exposing Insider Threats Using Eye Movement Biometrics. ACM Transactions on Privacy and Security. 2016;19(1):1:1–1:31. doi: 10.1145/2904018 [DOI] [Google Scholar]
  • 19.Zhang Y, Hu W, Xu W, Chou CT, Hu J. Continuous Authentication Using Eye Movement Response of Implicit Visual Stimuli. ACM Interactive Mobile Wearable Ubiquitous Technologies. 2018;1(4):177:1–177:22. [Google Scholar]
  • 20.Steil J, Hagestedt I, Huang MX, Bulling A. Privacy-aware Eye Tracking Using Differential Privacy. In: ACM Symposium on Eye Tracking Research & Applications. New York, NY, USA: ACM; 2019. p. 27:1–27:9.
  • 21.Narayanan A, Shmatikov V. Robust De-anonymization of Large Sparse Datasets. In: IEEE Symposium on Security and Privacy. New York, NY, USA: IEEE; 2008. p. 111–125.
  • 22.Dwork C, McSherry F, Nissim K, Smith A. Calibrating Noise to Sensitivity in Private Data Analysis. In: Halevi S, Rabin T, editors. Theory of Cryptography. Berlin, Heidelberg, Germany: Springer Berlin Heidelberg; 2006. p. 265–284. [Google Scholar]
  • 23.Dwork C. Differential Privacy. In: Automata, Languages and Programming. Berlin, Heidelberg, Germany: Springer Berlin Heidelberg; 2006. p. 1–12. [Google Scholar]
  • 24.Erlingsson U, Pihur V, Korolova A. RAPPOR: Randomized Aggregatable Privacy-Preserving Ordinal Response. In: ACM SIGSAC Conference on Computer and Communications Security. New York, NY, USA: ACM; 2014. p. 1054–1067.
  • 25.Ding B, Kulkarni J, Yekhanin S. Collecting Telemetry Data Privately. In: International Conference on Neural Information Processing Systems. Red Hook, NY, USA: Curran Associates Inc.; 2017. p. 3574–3583.
  • 26.Rastogi V, Nath S. Differentially Private Aggregation of Distributed Time-series with Transformation and Encryption. In: ACM SIGMOD International Conference on Management of Data. New York, NY, USA: ACM; 2010. p. 735–746.
  • 27.Günlü O, İşcan O. DCT Based Ring Oscillator Physical Unclonable Functions. In: IEEE International Conference on Acoustics, Speech and Signal Processing. New York, NY, USA: IEEE; 2014. p. 8198–8201.
  • 28.Günlü O, Kernetzky T, Işcan O, Sidorenko V, Kramer G, Schaefer RF. Secure and Reliable Key Agreement with Physical Unclonable Functions. Entropy. 2018;20(5). doi: 10.3390/e20050340 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Steil J, Koelle M, Heuten W, Boll S, Bulling A. PrivacEye: Privacy-preserving Head-mounted Eye Tracking Using Egocentric Scene Image and Eye Movement Features. In: ACM Symposium on Eye Tracking Research & Applications. New York, NY, USA: ACM; 2019. p. 26:1–26:10.
  • 30.Dwork C, Roth A. The Algorithmic Foundations of Differential Privacy. Foundations and Trends in Theoretical Computer Science. 2014;9(3-4):211–407. doi: 10.1561/0400000042 [DOI] [Google Scholar]
  • 31.Liebling DJ, Preibusch S. Privacy Considerations for a Pervasive Eye Tracking World. In: ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication. New York, NY, USA: ACM; 2014. p. 1169–1177.
  • 32.John B, Koppal S, Jain E. EyeVEIL: Degrading Iris Authentication in Eye Tracking Headsets. In: ACM Symposium on Eye Tracking Research & Applications. New York, NY, USA: ACM; 2019. p. 37:1–37:5.
  • 33.Bozkir E, Ünal AB, Akgün M, Kasneci E, Pfeifer N. Privacy Preserving Gaze Estimation Using Synthetic Images via a Randomized Encoding Based Framework. In: ACM Symposium on Eye Tracking Research and Applications. New York, NY, USA: ACM; 2020. p. 21:1–21:5.
  • 34.Chaudhary AK, Pelz JB. Privacy-Preserving Eye Videos Using Rubber Sheet Model. In: ACM Symposium on Eye Tracking Research and Applications. New York, NY, USA: ACM; 2020. p. 22:1–22:5.
  • 35.David-John B, Hosfelt D, Butler K, Jain E. A Privacy-preserving Approach to Streaming Eye-tracking Data. IEEE Transactions on Visualization and Computer Graphics. 2021;27(5):2555–2565. doi: 10.1109/TVCG.2021.3067787 [DOI] [PubMed] [Google Scholar]
  • 36.Li J, Chowdhury AR, Fawaz K, Kim Y. Kalϵido: Real-Time Privacy Control for Eye-Tracking Systems. In: USENIX Security Symposium. Berkeley, CA, USA: USENIX Association; 2021.
  • 37.Liu A, Xia L, Duchowski A, Bailey R, Holmqvist K, Jain E. Differential Privacy for Eye-tracking Data. In: ACM Symposium on Eye Tracking Research & Applications. New York, NY, USA: ACM; 2019. p. 28:1–28:10.
  • 38.Kifer D, Machanavajjhala A. Pufferfish: A Framework for Mathematical Privacy Definitions. ACM Transactions on Database Systems. 2014;39(1). doi: 10.1145/2514689 [DOI] [Google Scholar]
  • 39.Raval N, Machanavajjhala A, Pan J. Olympus: Sensor Privacy through Utility Aware Obfuscation. Proceedings on Privacy Enhancing Technologies. 2018;2019:5–25. doi: 10.2478/popets-2019-0002 [DOI] [Google Scholar]
  • 40.Cao Y, Yoshikawa M, Xiao Y, Xiong L. Quantifying Differential Privacy under Temporal Correlations. In: IEEE International Conference on Data Engineering. New York, NY, USA: IEEE; 2017. p. 821–832. [DOI] [PMC free article] [PubMed]
  • 41.Cao Y, Yoshikawa M, Xiao Y, Xiong L. Quantifying Differential Privacy in Continuous Data Release Under Temporal Correlations. IEEE Transactions on Knowledge and Data Engineering. 2019;31(7):1281–1295. doi: 10.1109/TKDE.2018.2824328 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.McSherry FD. Privacy Integrated Queries: An Extensible Platform for Privacy-preserving Data Analysis. In: ACM SIGMOD International Conference on Management of Data. New York, NY, USA: ACM; 2009. p. 19–30.
  • 43.Zhao J, Zhang J, Poor HV. Dependent Differential Privacy for Correlated Data. In: IEEE Globecom Workshops. New York, NY, USA: IEEE; 2017. p. 1–7.
  • 44.Kellaris G, Papadopoulos S. Practical Differential Privacy via Grouping and Smoothing. VLDB. 2013;6(5):301–312. doi: 10.14778/2535573.2488337 [DOI] [Google Scholar]
  • 45.Günlü O. Design and Analysis of Discrete Cosine Transform Based Ring Oscillator Physical Unclonable Functions [Master Thesis]. Technical University of Munich. Munich, Germany; 2013. [Google Scholar]
  • 46.Orekondy T, Schiele B, Fritz M. Towards a Visual Privacy Advisor: Understanding and Predicting Privacy Risks in Images. In: IEEE International Conference on Computer Vision. New York, NY, USA: IEEE; 2017. p. 3686–3695.

Decision Letter 0

Luca Citi

16 Apr 2021

PONE-D-21-07774

Differential privacy for eye tracking with temporal correlations

PLOS ONE

Dear Dr. Bozkir,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

In particular:

  • Make the motivation of this work clearer (why eye movement data are sensitive? why DP is vulnerable under temporal correlation? why it is reasonable to assume a trusted central server in this work?).

  • Please try to provide formal proofs of DP achieved by the proposed method. While this paper spans both biomedical engineering and security, most readers in the security domain would expect that claims about privacy are backed by formal proofs.

  • Consider adding related works as suggested by the reviewers.

  • Consider testing the methods on other datasets or provide arguments why this is not possible / not required.

  • Provide a comparison with existing methods in the literature, especially in terms of privacy rather than reconstruction of noisy data and its classification (or provide arguments why the latter covers the former).

  • Remove / reduce parts that appear to be disconnected form the rest of the paper or make the link clearer and justified.

Please submit your revised manuscript by May 31 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Luca Citi, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The paper “Differential privacy for eye tracking with temporal correlations” has proposed a novel data encoding technique based on differential privacy mechanism to preserve privacy of eye data for various VR/AR applications taking into account temporal correlation.

Specifically, the authors have extended the standard Laplacian Perturbation Algorithm (LPA) and Fourier Perturbation Algorithm (FPA) and have applied them for noise generation and dealing with the eye data correlation issue respectively.

Results and tables presented have significantly shown and proven chunk-based FPA (CFPA) and Difference- and chunk-based FPA (DCFPA), both as extended FPA proposed by authors are more efficient for privacy (data utilities) but provides low usability (data identification or classification) as compared to the standard FPA thus validating author’s claims on proposed methods and conclusions.

The paper is well written and clearly explained. Notwithstanding authors may consider the following minor suggestions:

• Spotted typo in abstract, “extend” instead of “extent”

• To further demonstrate the versatility of the proposed method, authors may consider expanding their evaluation to include other configurations that are not limited to MPIIDPEye and MPIIPrivacEye datasets. Alternatively, authors may give more justification to why proposed method evaluation was limited to MPIIDPEye and MPIIPrivacEye configurations other than just being the only non-commercial scientific purpose datasets.

Reviewer #2: The authors are studying an interesting topic of privacy-preserving eye tracking data release, which is important given the rapid advancement of VR and AR technologies. However, this paper has a lot room to improve. Here I list several essential weak points in the hope that the authors could provide a better version of this work.

W1. The motivation of this work needs to be clarified. For example, why eye movement data are sensitive? Why DP is vulnerable under temporal correlation? (also see W3) Why it is reasonable to assume a trusted central server in this work? (basically, LDP is perforable than central DP since LDP has less trust assumption)

W2. The technical depth of the proposal method is limited. The proposed Chunk-based and difference-based methods seem trivial to me. The privacy guarantee is not clear; especially, how they could guarantee DP under temporal correlations. The authors may want to provide the formal proofs.

W3. Important related works are missing. The following studies [a,b] demonstrate the vulnerability of DP under temporal correlations. The authors fail to acknowledge them and did not discuss whether or not the proposed methods in this work can address the vulnerability proven in [a,b].

[a] Quantifying Differential Privacy under Temporal Correlations, IEEE ICDE 2017.

[b] Quantifying Differential Privacy in Continuous Data Release Under Temporal Correlations. IEEE TKDE 2019.

Reviewer #3: Bozkir et al proposed a differential privacy method based on temporal signal processing methods to overcome possibly biomimetic information eavesdropping or information stealing. Their results focus on the correlation analysis of public available data in order to show different possible signal states. Then they apply a utility function in order to analyse how different techniques extended from the temporal signal processing methods will lead to less error in classifying the signal states when noise is added to the raw signal, thus implementing the differential privacy scheme.

The work is interesting and certainly has a timely contribution to this area. However, from the paper's current state, it is not clear weather there is significant benefits to privacy schemes of eye movement data because i) the authors do not provide a direct comparison with existing methods in the literature, and most of the evaluated techniques are based on simple extension of temporal signal processing methods, ii) the metrics used for evaluating the privacy technique are focused on reconstruction of noisy data and its classification, but not entirely on how efficient in terms of privacy the proposed techniques really are.

Moreover, in the last part of the paper, the authors spend a considerable amount of space on a totally different focus, which is classifying the different cohort features in the obtained datasets. This part has not a clear link to the differential privacy techniques and thus takes away a lot from the main message of the paper. I recommend the authors would make more space for evaluating further the proposed privacy methods directly.

However, I do think if the authors address the main points, this paper could lead to nice findings on how evaluating temporal signals correctly address the challenges in differential privacy.

Minor comments below:

- In: "Soon, the decrease in the cost of such devices might cause a mass consumption across different application domains such as gaming, entertainment, or education. " - How soon and a decrease from how much? Consider more precision in such statements.

- In: "As biometric contents can be retrieved from eye movements, it is important to protect them against adversarial attacks." - At this point the reader may not necessarily know what an adversarial attack is. Therefore, add a sentence here about what are adversarial attacks and why they are important.

- In: "To apply differential privacy to the eye movement data, we evaluate the standard Laplacian Perturbation Algorithm (LPA) [22] and Fourier Perturbation Algorithm (FPA) [25]. " - Why chose particularly those two algorithms? This must be motivated.

- In: "sing differential privacy, noise is added to the outcome of a function so that the outcome does not significantly change based on whether or not a randomly chosen individual participated in the dataset." - This sentence is confusing.

- In: "In addition, it is not possible to recognize the person accurately from eye movement data when the DCFPA is used." - Here the authors also show how the classification work, in the last part of the paper, may also not be a good result, which reinforces the need to either shrink it down or remove it completely.

- In: "Our methods and the current results form the state-of-the-art for differential privacy for aggregated eye movement feature signals." - This sentence is incomplete.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Aug 17;16(8):e0255979. doi: 10.1371/journal.pone.0255979.r002

Author response to Decision Letter 0


27 May 2021

We thank the Academic Editor and the Reviewers for their useful suggestions and comments that significantly helped to improve the paper, and provide our rebuttal letter below, following the order of reviewers’ comments. The main comments of the Academic Editor and the Reviewers are enumerated, and our responses are provided below the enumerated parts. In addition, we improved the readability of our manuscript. Please also see the revised version of our manuscript for the changes made. We hope that all concerns have been addressed in a satisfying way.

Response to the Academic Editor:

1) Make the motivation of this work clearer (why eye movement data are sensitive? why DP is vulnerable under temporal correlation? why it is reasonable to assume a trusted central server in this work?).

Thank you very much for your request to clarify our motivations. Since it is possible to authenticate and identify individuals by using solely their eye movements data, they are considered as sensitive information. We motivate this better in the revised version of the “Introduction” section of our paper (Please refer to the first paragraph of the “Introduction”.). We also explain the vulnerability of the DP to the temporal correlations and the assumption of the trusted server in our rebuttal below under the points (10) and (11), respectively. We especially motivate the temporal correlation issue in our paper now much clearly also by including the works suggested by the Reviewer 2 (See the “Previous research” section of our revised manuscript.).

2) Please try to provide formal proofs of DP achieved by the proposed method. While this paper spans both biomedical engineering and security, most readers in the security domain would expect that claims about privacy are backed by formal proofs.

In the revised manuscript, we provide the formal proof of FPA, which is our proof and which corrects the mistaken proof in the literature, and discuss the reasons why the chunk- and difference-based methods preserve the privacy via parallel and sequential compositions in CFPA and DCFPA sections, respectively.

3) Consider adding related works as suggested by the reviewers.

We have added the related work suggested by the reviewers and discussed them along with our methods especially in the “Previous research” section. Please see point (13) below for more details.

4) Consider testing the methods on other datasets or provide arguments why this is not possible / not required.

We have evaluated our methods on two prominent, recent, and publicly available eye movement datasets that are obtained from VR/AR setups. As the decorrelation effect that is provided by the Fourier transformation, partitioning into chunks, and converting the data to difference (or subtracted) signals does not strongly depend on the datasets (due to the strong decorrelation efficiency of the Fourier transform), it is reasonable to say that testing the proposed methods on other datasets is not necessary from a theoretical perspective. For example, by partitioning the data into chunks, the query sensitivities that should be applied to achieve the differential privacy decrease automatically and it is valid for any data (i.e., for any probabilistic model for the data). One trade-off that might depend on the datasets is the required sizes of used chunks to achieve a target level of decorrelation effect. We have already discussed this fact in the “Methods” section of our paper. If very small sized chunks are used, the data might still be temporally correlated, which increases the actual privacy metric epsilon-prime. With larger chunk sizes, temporal correlations between the chunks decrease; however, then the query sensitivities increase. Therefore, when applying the proposed methods, this trade-off should be taken into consideration to decide a reasonable chunk size for a given dataset. This is essentially not directly related to the raw eye movement data, but to the feature extraction pipeline of aggregated eye movement features. For instance, in the MPIIDPEye and MPIIPrivacEye datasets, the sliding window size was selected as 0.5 and 1 seconds during the feature generation, respectively. If a larger sliding window is selected, the temporal correlations will be lower in the aggregated features. This shows that an empirical evaluation and decision should be made when selecting the chunk sizes to generate private data. Other than these points, the proposed methods are generic and could be applied to all datasets or data types. We motivate this insight in a better way in the updated manuscript.

5) Provide a comparison with existing methods in the literature, especially in terms of privacy rather than reconstruction of noisy data and its classification (or provide arguments why the latter covers the former).

We have already compared our results with the existing methods that apply differential privacy to the aggregated eye movement features, the standard Laplacian mechanism used to provide differential privacy, and the (corrected) FPA method that is suitable for time-series data, and we used the NMSE and classification accuracy metrics as our utility metric for comparisons. We clarify these below under point (14) in our rebuttal and strengthen our argumentations in our revised version of the manuscript.

6) Remove / reduce parts that appear to be disconnected form the rest of the paper or make the link clearer and justified.

Thank you very much for your suggestion! We have justified the reasons for using classification accuracy metric in our paper at the beginning of the “Results” section and under the point (16) below. We argue that the normalized mean square error (NMSE) metric is an analytically trackable utility metric to compare the original and private versions of the signals, whereas classification accuracies of different tasks (which is another utility metric that is difficult to track analytically) provides insights into the practical usability of the private eye movement signals. Such an evaluation scheme is also in line with the previous research in the field of privacy-preserving eye tracking (Steil et al. 2019 [20], David-John et al. 2021 [34] in the paper), as the main goal of these works is to make the persons unidentifiable while keeping the utility of the data as high as possible.

Thank you very much for your comments!

Response to the Reviewer #1:

Reviewer #1: The paper “Differential privacy for eye tracking with temporal correlations” has proposed a novel data encoding technique based on differential privacy mechanism to preserve privacy of eye data for various VR/AR applications taking into account temporal correlation. Specifically, the authors have extended the standard Laplacian Perturbation Algorithm (LPA) and Fourier Perturbation Algorithm (FPA) and have applied them for noise generation and dealing with the eye data correlation issue respectively.

Results and tables presented have significantly shown and proven chunk-based FPA (CFPA) and Difference- and chunk-based FPA (DCFPA), both as extended FPA proposed by authors are more efficient for privacy (data utilities) but provides low usability (data identification or classification) as compared to the standard FPA thus validating author’s claims on proposed methods and conclusions. The paper is well written and clearly explained. Notwithstanding authors may consider the following minor suggestions:

7) Spotted typo in abstract, “extend” instead of “extent”

We have updated the typo in the abstract from “extent” to “extend”. Thank you!

8) To further demonstrate the versatility of the proposed method, authors may consider expanding their evaluation to include other configurations that are not limited to MPIIDPEye and MPIIPrivacEye datasets. Alternatively, authors may give more justification to why proposed method evaluation was limited to MPIIDPEye and MPIIPrivacEye configurations other than just being the only non-commercial scientific purpose datasets.

We use the MPIIDPEye and MPIIPrivacEye datasets since these datasets are benchmarks dedicated to privacy-preserving eye tracking in the VR/AR area. Both datasets consist of aggregated and timely features related to eye fixations, saccades, blinks, and pupil diameters which are commonly used in VR/AR applications since they represent individual user visual behaviors. In addition, the state-of-the-art works are also evaluated using these datasets. That is why we evaluated our methods on these datasets. We clarify this in the “Datasets” section of our paper. Furthermore, as we keep the signal sizes small by partitioning the data into chunks, function sensitivities decrease as well as the amount of required noise to obtain differential privacy. This operation can be applied independent of the dataset as long as temporal signals are used. Therefore, as we also mention in the “Conclusion” section of our paper, our methods could be applied to other biometric related data as well.

Thank you very much for your comments!

Response to the Reviewer #2:

Reviewer #2: The authors are studying an interesting topic of privacy-preserving eye tracking data release, which is important given the rapid advancement of VR and AR technologies. However, this paper has a lot room to improve. Here I list several essential weak points in the hope that the authors could provide a better version of this work.

9) The motivation of this work needs to be clarified. For example, why eye movement data are sensitive?

Thank you very much for your request to clarify our motivations. Since it is possible to authenticate and identify individuals by using their eye movements data (because they have unique characteristics and they can leak private information such as whether an individual is going through depression, which can be detected by checking the eye-movement reaction time that is on average longer for a depressed person), they are considered as sensitive information. We motivate this in the “Introduction” section of our paper (Please refer to the first paragraph of the “Introduction”.).

10) Why DP is vulnerable under temporal correlation? (also see W3)

Thank you for your request to explain the vulnerability of DP mechanisms to correlations. As an example, we can consider that Alice has provided her eye movement feature signals. As illustrated for the publicly available datasets in our paper, such feature signals are highly correlated over time due to the nature of feature generation pipeline of eye movements (e.g., using windows sizes of 30 seconds with the step size of 0.5 and 1 seconds for MPIIDPEye and MPIIPrivacEye, respectively.). As Alice’s eye movement data at time t=1 and t=2 are almost same [due to high correlation], adding Laplacian noise component (as an example DP mechanism) to each raw data point independently will not hide the sensitive information because the average of two noisy data points will be very close to the original raw values due to the symmetry in the Laplacian noise (Please refer especially to [25, 42, 43] in the paper. Refer to “Steil et al. 2019 [20]” in the paper for the shortcomings of the standard differential privacy mechanisms on eye movements.). Taking these into consideration, in order to hide the sensitive information and provide differential privacy for eye movements, methods that take data correlation into consideration should be applied, such as our methods that aim at eliminating the correlations before applying any DP mechanism. We also clarify the motivation in the “Previous research” section.

11) Why it is reasonable to assume a trusted central server in this work? (basically, LDP is perforable than central DP since LDP has less trust assumption)

In most of the current VR/AR application use-cases, the user data, particularly eye tracking data, are collected by trusted entities and stored in a centralized fashion to model users’ behaviors to improve their product, as well as possibly due to the fact that VR/AR devices are not commonly available in the households. After having the complete data, various analyses are done such as user visual behaviors, gaze-guidance use-cases, foveated rendering, saliency estimation etc. by these trusted entities. That is why we opted for a global differential privacy setting by assuming a trusted server. While being out of scope of this work, local differential privacy is surely a valid assumption with the scenario that each user sends their data to the server after applying local differential privacy mechanisms locally. However, for the eye tracking data collected from VR/AR setups, one should also guarantee that each user applies calibration procedures or similar steps properly to obtain valid and high quality data, and there is already some research in that direction [1,2]. Overall, we foresee that both scenarios will go hand in hand in the near future regarding privacy-preservation since both setups are valid. For this work, we focus on the global differential privacy scenario that seems to fit well to the recent VR/AR applications. We clarify this at the end of the second paragraph of the “Introduction” section. Thank you!

12) The technical depth of the proposal method is limited. The proposed Chunk-based and difference-based methods seem trivial to me. The privacy guarantee is not clear; especially, how they could guarantee DP under temporal correlations. The authors may want to provide the formal proofs.

Thank you very much for your suggestion! We are very happy to provide further proofs, including a correction of the wrong result in the literature. In the revised manuscript, we provide the formal proof of FPA, which is our proof and which corrects the mistaken proof in the literature, and discuss the reasons why the chunk- and difference-based methods preserve the privacy via parallel and sequential compositions in CFPA and DCFPA sections, respectively.

13) Important related works are missing. The following studies [a,b] demonstrate the vulnerability of DP under temporal correlations. The authors fail to acknowledge them and did not discuss whether or not the proposed methods in this work can address the vulnerability proven in [a,b].

[a] Quantifying Differential Privacy under Temporal Correlations, IEEE ICDE 2017.

[b] Quantifying Differential Privacy in Continuous Data Release Under Temporal Correlations. IEEE TKDE 2019.

Thank you very much for pointing out the missing related works. We have included them in our revised manuscript. These works mainly focus on the privacy leakage of standard differential privacy mechanisms and efficient ways to calculate the leakage in polynomial time. As we acknowledge such leakage of standard mechanisms, we have proposed low-complexity methods to reduce the data correlations to achieve differential privacy mechanisms that do not leak extra privacy due to the temporal correlations. We clarify these in “Previous research”, “Chunk-based FPA”, and “Difference- and chunk-based FPA” sections. As illustrated in the “Correlation analysis” section, using our methods along with the FPA, temporal correlations reduce significantly thanks to the high decorrelation efficiency of the DFT and additional decorrelations that we provide with our methods. The privacy leakage due to the temporal correlations reduces significantly by using the methods proposed in [a] and [b] as well. However, our approach is different from the one in these papers as in these papers the extra privacy loss for fixed differential privacy mechanisms is quantified when temporal correlations exist. Instead, we try to reduce the correlations in advance so that our proposed methods do not suffer from the privacy leakage due to the temporal correlations, which allows the use of standard differential privacy mechanisms for any VR/AR dataset.

Thank you very much for your comments!

Response to the Reviewer #3:

Reviewer #3: Bozkir et al proposed a differential privacy method based on temporal signal processing methods to overcome possibly biomimetic information eavesdropping or information stealing. Their results focus on the correlation analysis of public available data in order to show different possible signal states. Then they apply a utility function in order to analyse how different techniques extended from the temporal signal processing methods will lead to less error in classifying the signal states when noise is added to the raw signal, thus implementing the differential privacy scheme. The work is interesting and certainly has a timely contribution to this area. However, from the paper's current state, it is not clear whether there are significant benefits to privacy schemes of eye movement data.

14) The authors do not provide a direct comparison with existing methods in the literature, and most of the evaluated techniques are based on simple extension of temporal signal processing methods.

Thank you very much for your request! As reported in our paper, we have compared our methods/results with the state-of-the-art works in the eye tracking domain (Steil et al. 2019a [20], Steil et al. 2019b [28] in the paper) in terms of document-type, gender, and scene privacy sensitivity classifications from a practical point of view. In addition, we also compare our results with the standard Laplacian mechanism used to provide DP, and the corrected version of the FPA in terms of the NMSE. The results show that we outperform both standard Laplacian mechanism and the FPA, the latter of which being suitable for time-series with temporal correlations. In addition, while hiding sensitive information such as personal identifiers and gender information with CFPA and DCFPA, document-type prediction and scene privacy sensitivity detection accuracies based solely on eye movements are close to the accuracies obtained from non-private mechanisms. Applying the FPA does not hide personal identifiers as we are able to show that the person identification task works quite accurately after the application of the FPA. The strengths of CFPA and especially DCFPA stem from the significant decorrelation of the eye movement signals. We have strengthened our manuscript discussing these in a clearer way. In addition, in the literature, there are some works which do not formally provide privacy, which we have discussed in the “Previous research” section. Such works rather apply simple mechanisms such as adding a pre-defined noise based on Gaussian distributions or spatial/temporal downsampling. Since the scope of such works are different compared to DP mechanisms that require mathematical privacy proofs, we have not directly compared their results with ours.

15) The metrics used for evaluating the privacy technique are focused on reconstruction of noisy data and its classification, but not entirely on how efficient in terms of privacy the proposed techniques really are.

In this work, we have focused on the epsilon-DP that is applied to eye movement feature signals. The reconstruction of eye movement signals is only one step in the proposed methods, which provide differential privacy. Therefore, the metric that we use to measure the privacy level is epsilon, which is the same for other differential privacy studies as well. In our study, the difference from the other traditional approaches is that observations of the eye movement signals are highly correlated. Therefore, we aim to decrease the correlation before applying differential privacy mechanisms to the data. As also stated in various places in our paper, if the data are correlated, epsilon-prime is the actual privacy metric that should be used in practice since epsilon-prime is the differential privacy level that can be achieved against an attacker that is informed about the data correlations. In order to keep the epsilon-prime close to the epsilon value itself, we apply various methods to decrease the data correlations. Thus, applying our methods the epsilon metric could be used along with the decorrelated data. We empirically show that data become decorrelated with the different sizes of chunks and difference (or subtraction) signals that are calculated. Overall, we discuss that there is a trade-off between the chunk size and remaining correlation, so the chunk size should be empirically decided for each dataset since the feature generation pipelines could affect the temporal correlations positively or negatively in terms of privacy. We strengthen our arguments especially in the “Discussion” section based on your suggestions.

16) Moreover, in the last part of the paper, the authors spend a considerable amount of space on a totally different focus, which is classifying the different cohort features in the obtained datasets. This part has not a clear link to the differential privacy techniques and thus takes away a lot from the main message of the paper. I recommend the authors would make more space for evaluating further the proposed privacy methods directly.

We clarify why we additionally used classification accuracies for evaluation in the beginning of the “Results” section. In general, low normalized mean square error (NMSE) between original and private signals is favorable as this implies less variation from the original signals and this metric is analytically trackable. However, in practice, the private signals should be usable, i.e., different tasks should be applicable accurately. For this purpose, aligning with the previous research on the privacy-preserving eye tracking, we have applied document-type prediction and scene privacy sensitivity detection tasks. In addition, again aligning with the previous work, we have analyzed whether gender prediction and person identification tasks could be applied accurately on the private signals, which is undesirable (David-John et al. 2021 [34], Steil et al. 2019 [20] in the paper). In general, while providing the differential privacy, it is important to hide the personal information in the data and keep the data usable in practice. That is why we have used two utility metrics for evaluation and comparison, namely the NMSE metric and the classification accuracies. We show that with CFPA and DCFPA, while we keep the NMSE and person identification accuracies low, classification accuracies of document-type and scene privacy sensitivities high. On the contrary, the FPA method has both higher NMSE and higher person identification accuracies, which implies it is still possible to infer significant amounts of personal information even though a high amount of noise is added. Since the decorrelation efficiency of the DCFPA is the highest, person identification accuracies are around the guessing probabilities for this method. Overall, we think that providing both metrics for evaluation and comparison gives a lot of insights about the privacy mechanisms and in line with the current literature. Therefore, we prefer keeping them in our paper with an improved discussion, especially in the “Results” and “Discussion” sections.

17) Minor comments:

- We have updated the statement on HMDs and removed the argumentation on the cost of the devices as the cost itself is not necessarily relevant to our work due to the central nature of the global DP model we used (See the first paragraph of the “Introduction” section.).

- We have clarified the statement on adversarial behaviors (See the end of first paragraph and the beginning of second paragraph in the “Introduction” section.).

- We have already explained why we have chosen the LPA and the FPA in the third paragraph of the “Introduction” section and also in the “Fourier Perturbation Algorithm (FPA)” section. The LPA is the standard Laplace mechanism of differential privacy and FPA is one suitable mechanism for correlated data due to high decorrelation efficiency of the Discrete Fourier Transform (DFT) that is used in that algorithm.

- The confusing sentence is updated.

- When the DCFPA is used, the classification accuracies of document-type or privacy sensitivity tasks are kept in acceptable and usable ranges. At the same time, it is not possible to classify the participant information using the eye movement data when this method is applied in both datasets. Not being able to classify the participants is indeed a good thing which shows that also from a classification perspective, the personal information that eye movement features are hidden. Previous work (Steil et al. 2019 [20] in the paper) also reports similar metrics for person identification. That is why, we think that it is quite relevant to our work and prefer to keep it in the paper.

- We have updated the incomplete sentence.

Thank you very much for your comments!

References:

[1] E. Bozkir, S. Eivazi, M. Akgün and E. Kasneci, “Eye Tracking Data Collection Protocol for VR for Remotely Located Subjects using Blockchain and Smart Contracts,” 2020 IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR), 2020, pp. 397-401, doi: 10.1109/AIVR50618.2020.00083.

[2] X. Ma, M. Cackett, L. Park, E. Chien and M. Naaman, “Web-Based VR Experiments Powered by the Crowd,” 2018 World Wide Web Conference (WWW), 2018, pp. 33-43, doi:10.1145/3178876.3186034.

Decision Letter 1

Luca Citi

7 Jul 2021

PONE-D-21-07774R1

Differential privacy for eye tracking with temporal correlations

PLOS ONE

Dear Dr. Bozkir,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that the paper has improved significantly and is quite close to meeting PLOS ONE’s publication criteria. We have a small number of outstanding points that we would like the authors to consider before publication. To speed up the process and in consideration of the reviewers' time, the paper won't be sent out to the reviewers again and instead a decision will be made by the editor only.

Please submit your revised manuscript by Aug 21 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A brief rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Luca Citi, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: (No Response)

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: (No Response)

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: (No Response)

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: (No Response)

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: I would agree with the acceptance of this paper since the authors nicely address my concerns with the previous version.

Reviewer #3: The authors have successfully addressed my comments. However, relevant information was found in the responses but not adequately included in the text. I recommend the authors to address the following points

- The introduction does not fully explain the work in a high level manner adequately. Some important motivation for parts of the work are still missing, namely a) the explanation about the motivation of the utility analysis is missing b) the lack of motivation for using the epsilon-DP as a privacy metric, even though used in other works, the reader needs to know why such metric.

- I recommend the authors to also think of a visual way to present their work, as it is not straightforward to understand how all the different privacy functions are used. I would also think of another way to present Algorithm 1, at the moment that algorithm seems to be not needed.

- In many answers in the response letter, the authors claim to have added more information in the discussion section. However, the additions made seem quite limited in terms of new information.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Aug 17;16(8):e0255979. doi: 10.1371/journal.pone.0255979.r004

Author response to Decision Letter 1


23 Jul 2021

We thank the Academic Editor and the Reviewers for evaluating our revised manuscript along with our rebuttal and for their helpful suggestions that helped to improve the manuscript. As further points are requested only by Reviewer 3, in our rebuttal, we address the points raised by the Reviewer 3. We hope that all concerns have been addressed in a satisfying way.

Response to Reviewer 3:

1) The introduction does not fully explain the work in a high level manner adequately. Some important motivation for parts of the work are still missing, namely a) the explanation about the motivation of the utility analysis is missing b) the lack of motivation for using the epsilon-DP as a privacy metric, even though used in other works, the reader needs to know why such metric.

Thank you very much for the suggestions.

a) The utility analysis based on the metric normalized mean square error (NMSE) shows the trend of divergence of noisy aggregated signals (i.e., differentially private signals) from the original signals and this metric is analytically trackable. We have stated this in the first paragraphs of “Results” and “Classification accuracy results” sections. To make this point clearer upfront, we have added this motivation to the last paragraph of “Introduction” section (Before “Previous research”) so that the reader can grasp the high level motivation before diving into technical details.

b) As requested, we have further explained the differential privacy and motivated its usage for eye movements in the second paragraph of the “Introduction” section. More technical details are already available in “Materials and methods” section.

2) I recommend the authors to also think of a visual way to present their work, as it is not straightforward to understand how all the different privacy functions are used. I would also think of another way to present Algorithm 1, at the moment that algorithm seems to be not needed.

Thank you very much for the suggestions. We have added a figure to visually depict FPA (instead of the algorithmic representation of it), since it is a fundamental approach that we use throughout the manuscript. In addition, we have added a visual representation combining the CFPA and DCFPA as suggested. These figures could be seen as Figures 1 and 2, respectively.

3) In many answers in the response letter, the authors claim to have added more information in the discussion section. However, the additions made seem quite limited in terms of new information.

Thank you very much for your request. We discuss our findings in “Discussion” section as well as providing their implications. While our “Results” section focuses on experimental evaluations, their implications and comparisons with our initial expectations are discussed in “Discussion”. To deepen the understanding based on our findings and comparisons, we have added further points to our “Discussion” section, especially to the first paragraph and strengthened the section by analyzing our findings in a more compact manner, especially by giving more details about the points mentioned in the previous revision and our previous rebuttal.

Thank you very much for your suggestions. We hope that all concerns have been addressed in a satisfying manner.

Attachment

Submitted filename: Response to Reviewers.pdf

Decision Letter 2

Luca Citi

28 Jul 2021

Differential privacy for eye tracking with temporal correlations

PONE-D-21-07774R2

Dear Dr. Bozkir,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Luca Citi, PhD

Academic Editor

PLOS ONE

Acceptance letter

Luca Citi

5 Aug 2021

PONE-D-21-07774R2

Differential privacy for eye tracking with temporal correlations

Dear Dr. Bozkir:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Luca Citi

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Document type classification results without majority voting for the MPIIDPEye dataset.

    (PDF)

    S2 Table. Gender classification results without majority voting for the MPIIDPEye dataset.

    (PDF)

    S3 Table. Person identification results without majority voting for the MPIIDPEye dataset.

    (PDF)

    S4 Table. Person identification results without majority voting for MPIIPrivacEye dataset.

    (PDF)

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Data Availability Statement

    Relevant data files are provided via following url: https://atreus.informatik.uni-tuebingen.de/bozkir/dp_eye_tracking.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES