Abstract
With the compelling popularity of integrated sensing and communication (ISAC), Wi-Fi sensing has drawn increasing attention in recent years. Starting from 2010, Wi-Fi channel state information (CSI)-based wireless sensing has enabled various exciting applications such as indoor localization, target imaging, activity recognition, and vital sign monitoring. In this paper, we retrospect the latest achievements of Wi-Fi sensing using commodity-off-the-shelf (COTS) devices from the past 5 years in detail. Specifically, this paper first presents the background of the CSI signal and related sensing models. Then, recent studies are categorized from two perspectives, i.e., according to their application scenario diversity and the corresponding sensing methodology difference, respectively. Next, this paper points out the challenges faced by Wi-Fi sensing, including domain dependency and sensing range limitation. Finally, three imperative research directions are highlighted, which are critical for realizing more ubiquitous and practical Wi-Fi sensing in real-life applications.
Keywords: Wi-Fi sensing, CSI, commodity-off-the-shelf, integrated sensing and communication
1. Introduction
The demand for ubiquitous internet connection has catalyzed the vast deployment of Wi-Fi infrastructures over the past decades, making Wi-Fi signal available almost everywhere. With the rapid progress of wireless communication and signal processing techniques, researchers have successfully reused Wi-Fi as a sensing platform beyond its traditional use as a pure communication medium, which further gives birth to the idea of integrated sensing and communication (ISAC) with Wi-Fi [1,2,3]. After years of persistent research, Wi-Fi sensing is drawing huge attention from both academia and industry [4]. Both communities recognize ISAC as a compelling technology capable of improving spectrum efficiency and reducing the hardware cost [5]. It is worth mentioning that, starting from 2020, the IEEE 802.11 working group established an IEEE 802.11bf standardization group for encompassing wireless sensing within the new version of 802.11 standard, turning Wi-Fi sensing into reality.
The basic rational behind Wi-Fi sensing is quite straightforward [6]. When wireless signal propagates from the transmitter to the receiver through multiple paths, a phenomenon called multi-path effect occurs, whereby the superimposed receiving signal intrinsically contains the signal component reflected or diffracted by the sensing target. Therefore, by analyzing the target “modulated” receiving signals, researchers can recover the rich information regarding the target, such as location and activity. Compared with classic sensor-based and vision-based sensing paradigms, Wi-Fi wireless sensing has the advantages of low-cost ubiquity, wide coverage, non-intrusiveness, and privacy-protection. Due to its appealing superiority, numerous Wi-Fi sensing applications have been developed, ranging from coarse-grained motion detection [7] and activity recognition [8] to fine-grained localization [9] and breath monitoring [10].
Inspired by existing survey papers [11,12,13,14,15], this paper investigates the thrilling achievements made within the last 5 years and presents an in-depth analysis of these sensing systems, aiming to facilitate further research in the Wi-Fi sensing field. This paper first divides existing works according to different application scenarios, including localization and tracking, activity recognition, vital sign monitoring, and target imaging. For each category, both application-specific problems and solutions are compared and summarized. Then, this paper further classifies recent studies based on the methodology adopted, whether it is model-based, handcrafted pattern extraction-based, or deep learning-based, pointing out the pros and cons of each method. Furthermore, this paper highlights the remaining challenges of current works such as generalization issues and large-scale perception. Future research directions and the need for further study are discussed in the end. The main contributions of this work are summarized as follows:
To the best of our knowledge, this is the latest comprehensive survey in the Wi-Fi sensing field, covering the greatest and most recent progresses made over the past 5 years.
We categorize existing studies from two distinct perspectives, i.e., application-based and methodology-based, and present an in-depth analysis of recent works.
We highlight the key challenges encountered in existing studies and present a thorough discussion about three promising research directions for Wi-Fi sensing.
The rest of this paper is organized as follows. In Section 2, we briefly introduce the concept of channel state information (CSI) and explain several popular sensing models. In Section 3, we classify state-of-the-art works with regard to two criteria, i.e., application variety and methodology difference. Practical limitations and challenges are analyzed in Section 4. In Section 5, a detailed discussion about future trends in Wi-Fi sensing is provided. Finally, we conclude this article in Section 6.
2. Preliminary
Before analyzing Wi-Fi sensing, we briefly introduce the necessary background of channel state information (CSI) and several general signal sensing models.
2.1. Channel State Information
Serving as a key metric of a communication system, CSI depicts how a signal propagates through a wireless channel. Indeed, a wireless communication channel can be defined as follows:
| Y = H × X + N | (1) |
where X and Y are the transmitted and received signal, respectively. H is the channel matrix representing CSI and N denotes the channel noise.
In a typical indoor environment, shown in Figure 1, a signal sent by the transmitter (Tx) travels through multiple paths before arriving at the receiver (Rx), which is also known as the multi-path effect. Therefore, assuming there are L different paths, the wireless channel H can be mathematically expressed as channel impulse response (CIR) [6], as follows:
| (2) |
where , , and are the complex amplitude attenuation, phase shift, and propagation time delay of the i-th path, respectively; is the Dirac delta function. Each impulse in the summation of Equation (2) represents a delayed multi-path component, multiplied by its corresponding amplitude and phase variation.
Figure 1.
Typical indoor multi-path Wi-Fi propagation.
As shown in Figure 1, when a person moves inside the signal zone, the human body will inevitably alter the specific propagation path, thus changing the CIR. Hence, the underlying principle of wireless sensing is analyzing human-induced channel variation. However, CIR cannot be precisely measured with commodity Wi-Fi devices, especially given the limited bandwidth of Wi-Fi. Fortunately, with the adoption of the orthogonal frequency division multiplex (OFDM) technique in present IEEE 802.11 standard, researchers resorted to studying channel frequency response (CFR), an equivalent channel representation of CIR in the frequency domain.
| (3) |
where and represent the amplitude–frequency and phase–frequency response of CFR, respectively. With proper driver modifications, researchers have been able to obtain an OFDM-based sampling version of CFR with a commercial-off-the-shelf (COTS) Wi-Fi network interface card (NIC) since 2010 [16,17], greatly prompting the development of Wi-Fi sensing [12]. To be specific, the extracted CFR depicts the amplitude and phase of different subcarriers, which can be expressed as follows:
| (4) |
where is the CFR sampled at the i-th subcarrier with the central frequency of . In fact, the CSI data used in most research papers refers directly to the definition given by Equation (4), i.e., a sampled version of CFR at the granularity of a subcarrier level.
Generally speaking, this sampled CFR lays the foundations for advanced Wi-Fi sensing, paving the way for the feasibility of various modern applications. CSI data contains rich information on signal propagation, and we will use CSI to simply signify the raw Wi-Fi data for brevity in the following part.
2.2. Signal Sensing Models
2.2.1. Fresnel Zone-Based Reflection Model
Taking one pair of the Tx-Rxlink as an example, Fresnel zones are concentric ellipses with two foci corresponding to the Tx and Rx, as P1 and P2, shown in Figure 2. For a given radio length , the n-th Fresnel zone boundary containing n ellipses can be defined as follows:
| (5) |
where is a point on the n-th Fresnel zone boundary. The n-th Fresnel zone refers to the elliptic annulus between the (n−1)-th and n-th ellipse boundary, while the innermost ellipse is called the first Fresnel zone (FFZ). Equation (5) indicates that the path length of the signal reflected through the n-th Fresnel zone boundary is is longer than that of the Line-of-Sight (LOS) path, i.e., .
Figure 2.
Geometry of Fresnel zone reflection sensing [18].
The Fresnel zone-based reflection model [18] characterizes how the amplitude and phase of CSI change when a target moves outside the FFZ. The key property of the reflection sensing model is when a target moves across a series of Fresnel zone boundaries, CSI amplitude and phase will show a continuous sinusoidal-like pattern, which can be utilized for sensing applications such as respiration and walking direction detection [19].
2.2.2. Fresnel Zone-Based Diffraction Model
According to the RF propagation theory, more than 70% of the signal energy is transferred via the FFZ. Therefore, when a target moves inside the FFZ, signal diffraction becomes more important and dominates the received signal variation. As shown in Figure 3, the Fresnel zone-based diffraction model [20] depicts how the amplitude and phase of CSI change when a target moves inside the FFZ. The key property is when sensing activity inside the FFZ, the CSI amplitude variation will show different shapes, be it either monotonically decreased or non-monotonous, “W”, depending on the target size. Apart from respiration monitoring, the diffraction sensing model has also been proved effective for recognizing exercise and daily exercise [8].
Figure 3.
Geometry of Fresnel zone diffraction sensing [20].
2.2.3. Scattering Sensing Model
One main limitation of the previous models is that the simple reflection or diffraction assumption may not hold true when considering complex target motions, in cases where signals are scattered from multiple human body parts. Different from the Fresnel zone-based model, the scattering sensing model treats all objects as scatters, taking into account all multi-paths together. As marked as red circles in Figure 4, intuitively, the scattering model considers each scatter point as a virtual Tx, e.g., the static walls, and the arm and leg of the moving human. Given numerous multi-paths are considered, the scattering model is in fact a statistical model generally applicable to complex indoor scenarios. The scattering sensing model has been adopted in various speed-oriented tasks [21,22], achieving robust performance even with non-line-of-sight (NLOS) occlusion.
Figure 4.
Signal scattering sensing model.
3. Wi-Fi Sensing
Serving as a key property in future wireless systems, Wi-Fi sensing has enabled various important applications. In this section, we categorize recent works based on two aspects, i.e., whether they are application-oriented or methodology-oriented. Since there are quite a few references in this section, for the reader’s convenience, we provide an index of all mentioned references in Table A1 of Appendix A.
3.1. Wi-Fi Sensing Applications
In this section, we divide the related works into seven categories, i.e., presence detection, gait recognition, gesture recognition, activity recognition, localization and tracking, vital sign monitoring, and pose construction and imaging, as shown in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7. In each table, “Application” implies detailed application demand, “User number” signifies the number of sensing targets supported by the study, “Device type” indicates the specific sensing equipment used, and “NLOS” shows whether the sensing system can work in a non-line-of-sight scenario or not.
Table 1.
Presence detection.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2022 | WiCPD [23] | In-car child presence detection | 96.56–100% real-time detection rate | 1 | NXP Wi-Fi chipset | Y |
| 2023 | Hu et al. [24] | Proximity detection | 95% and 99% true positive rate for distance-based and room-based detection | 1 | NXP Wi-Fi chipset | Y |
| 2024 | Zhu et al. [25] | Human and non-human differentiation | 95.57% average accuracy | 1 human or pet | COTS device | Y |
| 2024 | WI-MOID [26] | Edge device-based human and non-human differentiation | 97.34% accuracy and 1.75% false alarm rate | 1 human or non-human subject | Wi-Fi edge device | Y |
Presence detection. Presence detection determines whether a target exists or not within the sensing area and serves as the prerequisite for further sensing tasks. Target presence detection could enable many modern applications, such as security systems and smart homes. Although usually included as a detector module in most studies, there have been some new applications based on presence detection. As shown in Table 1, WiCPD [23] studied child presence detection in a smart car scenario, preventing potential harm to children if left alone in a vehicle. Hu et al. [24] considered target location relative to the sensing device, supporting more intelligent control systems using this area-aware context. In addition, Zhu et al. [25] and WI-MOID [26] further differentiated human from non-human targets to mitigate influence from unwanted objects, avoiding unnecessary false alarms.
Table 2.
Gait recognition.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2021 | GaitSense [27] | Gait-based human identification | 93.2% for 5 users and 76.2% for 11 users | 11 | Intel 5300 | N |
| 2021 | GaitWay [28] | Gait speed estimation | 0.12 m median error | 1 | Intel 5300 | Y |
| 2022 | CAUTION [29] | Gait-based human authentication | 93.06 average accuracy | 15 | TP-Link N750 router | N |
| 2022 | Wi-PIGR [30] | Gait recognition | 93.5% for single user and 77.15% for 50 users | 1–50 | Intel 5300 | N |
| 2023 | Auto-Fi [31] | Gesture and gait recognition | 86.83% for gesture; 79.61% for gait | 1 | Atheros chipset | N |
| 2023 | GaitFi [32] | Gait recognition | 94.2% accuracy | 12 | TP-Link N750 router | N |
| 2024 | Wi-Diag [33] | Multi-subject abnormal gait diagnosis | 87.77% average accuracy | 4 | Intel 5300 | N |
Gait recognition. Gait, a unique biomarker, refers to the distinctive walking character of different people and has been used for human identification and authentication applications. Early gait sensing works usually required users to walk on fixed trajectories within restricted areas, while recent studies, e.g., GaitSense [27], GaitWay [28], and Wi-PIGR [30], aimed for path independent gait recognition where users can walk along arbitrary paths even in a through-the-wall scenario. In addition, CAUTION [29], Auto-Fi [31], and GaitFi [32] tried to realize robust gait recognition with limited training data, while Wi-Diag [33] further studied more challenging multi-human recognition problems. As depicted in Table 2, all these works greatly contribute to more ubiquitous gait-based sensing applications.
Table 3.
Gesture recognition.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2021 | Kang et al. [34] | Gesture recognition | 3–12.7% improvement | 1 | Widar Dataset | N |
| 2021 | WiGesture [35] | Gesture recognition | 92.8–94.5% accuracy | 1 | Intel 5300 | N |
| 2022 | HandGest [36] | Handwriting recognition | 95% accuracy | 1 | Intel 5300 | N |
| 2022 | DPSense-WiGesture [37] | Gesture recognition | 94% average accuracy | 1 | Intel 5300 | N |
| 2022 | Niu et al. [38] | Gesture recognition | 96% accuracy | 1 | Intel 5300 | Y |
| 2022 | Widar 3.0 [39] | Cross-domain gesture recognition | 92.7% in-domain and 82.6–92.4% cross-domain accuracy | 1 | Intel 5300 | N |
| 2022 | WiFine [40] | Gesture recognition | 96.03% accuracy in 0.19 s | 1 | Raspberry Pi 4B | N |
| 2023 | UniFi [41] | Gesture recognition | 99% and 90–98% accuracy for in-domain and cross-domain recognition | 1 | Widar dataset | N |
| 2023 | WiTransformer [42] | Gesture recognition | 86.16% accuracy | 1 | Widar dataset | N |
| 2024 | AirFi [43] | Gesture recognition | 90% accuracy | 1 | TP-Link N750 router | N |
| 2024 | WiCGesture [44] | Continuous gesture recognition | 89.6% for digits and 88.3% for Greek letters | 1 | Intel 5300 | N |
Gesture recognition. Wireless gesture recognition has emerged as an important part of modern human computer interaction, enabling wide applications including smart home control and virtual reality. Previous studies tried to learn the intricate pattern between signal variation and human gesture under the one-to-one mapping assumption. However, this assumption does not hold, since the received signal is highly dependent on the relative location and orientation of users, as proven by the Fresnel reflection model [18]. Thus, recent works mainly focused on realizing a position-independent robust gesture recognition system, as illustrated in Table 3. Kang et al. [34], Widar 3.0 [39], UniFi [41], WiTransformer [42], and AirFi [43] leverage various deep learning methods, e.g., adversarial learning, multi-view network, and few-shot learning, to realize a robust and efficient recognition. On the other hand, WiGesture [35], HandGest [36], DPSense-WiGesture [37], Niu et al. [38], and WiCGesture [44] attempted to extract distinct and consistent features from a hand-oriented perspective, realizing reliable and continuous recognition either through more fine-grained signal segmentation or signal quality assessment. In addition, WiFine [40] managed to realize real-time gesture recognition using low-end edge devices, e.g., Raspberry Pi. Overall, these methods bring Wi-Fi gesture recognition one step closer to more practical uses.
Table 4.
Activity recognition.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2020 | Wang et al. [45] | People counting and recognition | 86% average accuracy | 4 | COTS devices | N |
| 2021 | Ma et al. [46] | Activity recognition | 97% average accuracy | 1 | Intel 5300 | N |
| 2021 | MCBAR [47] | Activity recognition | 90% average accuracy | 1 | Atheros chipset | N |
| 2021 | WiMonitor [48] | Location and activity monitoring | Not applicable | 1 | Intel 5300 | Y |
| 2022 | DeFall [49] | Fall detection | 95% detection rate and 1.5% false alarm rate | 1 | Intel 5300 | Y |
| 2022 | Ding et al. [50] | Activity recognition | 96.85% average accuracy | 1 | Intel 5300 | N |
| 2022 | EfficientFi [51] | Activity recognition | 98% accuracy | 1 | TP-Link N750 router | N |
| 2022 | TOSS [52] | Activity recognition | 82.69% average accuracy | 1 | Intel 5300 | N |
| 2023 | FallDar [53] | Fall detection | 5.7% false alarm rate and 3.4% missed alarm rate | 1 | Intel 5300 | Y |
| 2023 | SHARP [54] | Activity recognition | 95% average accuracy | 1 | ASUS RT-AC86U router | N |
| 2023 | Liu et al. [55] | Moving receiver-based activity recognition | 10°, 1 cm and 98% accuracy for direction, displacement, and activity estimation | 1 | COTS WiFi 6 device | N |
| 2023 | WiCross [56] | Target passing detection | 95% accuracy | 1 | Intel 5300 | N |
| 2024 | i-Sample [57] | Activity recognition | 10% accuracy gain | 1 | Intel 5300 | N |
| 2024 | MaskFi [58] | Activity recognition | 97.61% average accuracy | 1 | TP-Link N750 router | N |
| 2024 | MetaFormer [59] | Activity recognition | Improved accuracy in various cross-domain scenarios | 1 | SiFi, Widar, Wiar datasets | N |
| 2024 | SAT [60] | Activity recognition | Improved accuracy and robustness | 1 | Intel 5300 | N |
| 2024 | SecureSense [61] | Activity recognition under adversarial attack | Robust performance under various attacks | 1 | TP-Link N750 router | N |
| 2024 | Luo et al. [62] | Activity recognition | 98.78% accuracy | 1 | UT-HAR dataset | N |
| 2024 | WiSMLF [63] | Activity recognition | 92% average accuracy | 1 | Intel 5300 | N |
Activity recognition. Wi-Fi-based human activity recognition (HAR) has become the most studied research topic over the past years, covering many applications including people counting [45], fall detection [49,53], door-passing detection [56], and daily activities. Table 4 shows the summary of recent HAR works. Most works tried to address performance degradation due to location, person, and environment dynamics, also known as domain-dependent problems [46,47,50,52,54,57,58,59,62,63]. In addition, WiMonitor [48] studied continuous long-term human activity monitoring, capturing user information such as location change, activity intensity, and time. Moreover, EfficientFi [51] considered the signal transfer-induced communication problem in a large-scale sensing scenario, providing a cloud-enabled solution with efficient CSI compression, while SAT [60] and SecureSense [61] proposed robust sensing schemes under various adversarial attacks. Liu et al. [55] proposed a dynamic Fresnel zone sensing model using a moving receiver such as a smartphone, filling the gap in existing fixed-location transceivers.
Table 5.
Localization and tracking.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2022 | Niu et al. [64] | Velocity estimation-based tracing | 9.38 cm/s, 13.42° and 31.08 cm median error in speed, heading and location estimation | 1 | Intel 5300 | Y |
| 2023 | WiTraj [65] | Human walking tracking | 2.5% median tracking error | 1 | Intel 5300 | N |
| 2024 | FewSense [66] | Tracking | 34 cm median error | 1 | Intel 5300 | N |
| 2023 | Zhang et al. [67] | Multi-person localization | Sub-centimeter accuracy | 1–3 | COTS WiFi device + IRS | N |
| 2024 | Zhang et al. [68] | Passive localization | 0.11 m average error | 1 | VNA | N |
| 2022 | Fan et al. [69] | Moving direction estimation | 6.9° median error for moving direction estimation; 16.6° mean error for rotation angle estimation | 1 | Atheros chipset | Y |
| 2022 | Wi-Drone [70] | Tracking-based indoor drone flight control | 26.1 cm average location accuracy and 3.8° rotation accuracy | 1 | AR9580 NICs | N |
Localization and tracking. Due to the limited channel bandwidth and antenna number of COTS Wi-Fi devices, there have not been many studies on Wi-Fi-based localization and tracking, as shown in Table 5. Recent works tried to improve tracking performance through more accurate target velocity estimations using a moving-induced Doppler Frequency Shift (DFS). Niu et al. [64] optimized velocity estimation by devising a dynamic selection scheme, which can choose the optimal set of receivers for tracking. To better track human walking, WiTraj [65] intelligently combined multi-view information provided by different receivers and differentiated walking with in-place activity to avoid tracking error accumulation. FewSense [66] creatively fused phase and information for better DFS estimation, achieving high accuracy even with fewer CSI samples. In addition to these works, Zhang et al. [67,68] achieved sub-centimeter localization accuracy using the intelligent reflecting surface (IRS) technique. By constructing an IRS, researchers can modulate the spatial distribution of the Wi-Fi signal, improving the spatial resolution of Wi-Fi localization. While promising, their current prototype systems are realized using a vector network analyzer (VNA), requiring further study with a COTS device. Apart from the device-free tracking mentioned above, Fan et al. [69] and Wi-Drone [70] studied device-based tracking applications. Fan et al. [69] obtained accurate moving direction and in-place rotation angle estimation using a single access point, while Wi-Drone [70] realized the first Wi-Fi tracking-based indoor drone flight control system, providing promising possible solutions for indoor localization and navigation.
Table 6.
Vital sign monitoring.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2020 | MultiSense [71] | Multi-person respiration sensing | 0.73 bpm mean error | 4 | Intel 5300 | Y |
| 2021 | SMARS [72] | Breath estimation and sleep stage recognition | 0.47 bpm median error and 88% accuracy | 1 | Atheros chipset | Y |
| 2021 | WiFi-Sleep [73] | Sleep stage monitoring | 81.8% accuracy | 1 | Intel 5300 | N |
| 2021 | WiPhone [74] | Respiration monitoring | 0.31 bpm average error | 1 | ASUS RT-AC86U router and Google Nexus 5 smartphone | Y |
| 2022 | ResFi [75] | Respiration detection | 96.05% accuracy | 1 | ASUS RT-AC86U router | N |
| 2024 | Xie et al. [76] | Respiration sensing with interfering individual | 32% mean absolute error reduction | 1 | VNA or Intel 5300 | N |
Vital sign monitoring. Vital signs play a crucial role in monitoring people’s health and well-being, providing useful information for early prediction and interference with potential diseases. As shown in Table 6, CSI-based vital sign detection mainly focused on respiration estimation. MultiSense [71] studied the multi-person respiration sensing problem, while SMARS [72] and WiFi-Sleep [73] integrated breath monitoring into users’ sleep quality assessment. WiPhone [74] presented a smartphone-based sensing system, achieving robust performance in NLOS scenarios. Xie et al. [76] addressed the motion interference from nearby individuals, bringing respiration monitoring closer to practical application.
Table 7.
Pose construction and imaging.
| Year | Reference | Application | Performance | User Number | Device Type | NLOS |
|---|---|---|---|---|---|---|
| 2020 | WiPose [77] | Pose construction | 2.83 cm average error | 1 | Intel 5300 | N |
| 2020 | WiSIA [78] | Target imaging | Not applicable | 1 | Intel 5300 | N |
| 2022 | GoPose [79] | 3D human pose estimation | 4.7 cm accuracy | 1 or 2 | Intel 5300 | Y |
| 2022 | Wiffract [80] | Still object imaging | 86.7% letter reading accuracy | 1 | Intel 5300 | Y |
| 2023 | MetaFi++ [81] | Pose estimation | 97.3% for PCK@50 | 1 | TP-Link N750 router | N |
| 2023 | WiMeasure [82] | Object size measurement | 2.6 mm median error | 1 | Intel 5300 | N |
| 2024 | PowerSkel [83] | Pose estimation | 96.27% for PCK@50 | 1 | ESP 32 IoT SoC | N |
| 2024 | WiProfile [84] | 2D target Profiling | 1 cm median absolute error | 1 target with proper size range | Intel 5300 | N |
Pose construction and imaging. Wi-Fi-based pose estimation and target imaging provides a complementary solution to traditional camera-based perception. As listed in Table 7, WiPose [77], GoPose [79], MetaFi++ [81], and PowerSkel [83] proposed different 3D human skeleton construction frameworks, while WiSIA [78], Wiffract [80], and WiProfile [84] further investigated how to recover target images with Wi-Fi signals. Alternatively, WiMeasure [82] realized millimeter-level high-precision target size measurements, making up for a missing piece of Wi-Fi sensing. It should be noted that in order to achieve fine-grained imaging, the deployment of a high sampling rate and even a customized antenna are usually required, as shown in the subsequent tables. Therefore, Wi-Fi imaging is only applicable for specific application scenarios for the time being.
3.2. Wi-Fi Sensing Methodologies
In this section, we divide the related works into three categories, i.e., model-based sensing, hand-crafted statistical pattern extraction-based sensing, and automatic deep pattern extraction-based sensing, as shown from Table 8, Table 9 and Table 10. In each table, “Methodology” briefly describes the specific method adopted, and “Base signal” refers to the sensing signal constructed with raw CSI, including autocorrelation function (ACF), power spectrum density (PSD), Doppler frequency shift (DFS), body-coordinate velocity profile (BVP), and so on. In addition, “Setting” specifies the signal sampling rate required, the number of Tx-Rx pair used, and certain device settings used for the system implementation and performance evaluation.
Model-based sensing. Since model-based sensing methods have the clear advantage of interpretability, researchers have developed several models for describing the physical relationship between CSI variation and target behavior, detailed in Section 2. As shown in Table 8, the scattering model has been widely used for velocity and periodic pattern extraction [28,49,72], while the diffraction model has been adopted in near-the-LOS scenarios, i.e., within FFZ, for fine-grained sensing tasks [56,80,82,84]. Although less prevalent in Table 8 [55], the Fresnel zone-based reflection model is in fact the most used model. The reflection model is commonly implicitly incorporated in various sensing systems for quantitatively analyzing signal variations and identifying sensing limitations, thus guiding the implementation of more stable and reliable sensing systems [85,86,87].
Table 8.
Model-based sensing.
| Year | Reference | Methodology | Performance | Base Signal | Sensing Range | Setting |
|---|---|---|---|---|---|---|
| 2021 | GaitWay [28] | Scattering model | 0.12 m median error | ACF of CSI | 20 m × 23 m | 1500 Hz; single pair of Tx-Rx |
| 2021 | SMARS [72] | Scattering model | 0.47 bpm median error and 88% accuracy | ACF of CSI | 10 m | 30 Hz; single pair of Tx-Rx |
| 2022 | DeFall [49] | Scattering model | 95% detection rate and 1.5% false alarm rate | ACF of CSI | Multi-room | 1500 Hz; single pair of Tx-Rx |
| 2022 | Wiffract [80] | Keller’s Geometrical Theory of Diffraction | 86.7% letter reading accuracy | Power of CSI | 1.5 m | Two pairs of Tx-Rx; two-dimensional RX grid synthesis |
| 2023 | Liu et al. [55] | Dynamic Fresnel zone model | 10°, 1 cm and 98% accuracy for direction, displacement and activity estimation | CSI | Single room | 100 Hz; single pair of Tx-Rx |
| 2023 | WiCross [56] | Diffraction model-based phase pattern extraction | 95% accuracy | CSI ratio | 1 m | 1000 Hz; single pair of Tx-Rx |
| 2023 | WiMeasure [82] | Diffraction model | 2.6 mm median error | CSI ratio | Near the LOS path | 500 Hz; three pairs of Tx-Rx |
| 2024 | WiProfile [84] | Diffraction effect-based profiling + inverse Fresnel transform | 1 cm median absolute error | CSI | 1.5 m × 1 m | 500 Hz; single pair of Tx-Rx; One reference receiving antenna connected to Rx via feeder line |
Hand-crafted statistical pattern extraction-based sensing. Derived from feature engineering in traditional machine learning processes, researchers have come up with various task-oriented feature extraction schemes, utilizing the in-depth analysis of activity characteristics and advanced signal processing techniques. As shown in Table 9, along with signal processing such as signal segmentation and signal energy estimation, statistical features, such as Doppler frequency shift and speed estimation, motion navigation primitive (MNP), dynamic phase vector (DPV) and motion rotation variable (MRV), have been derived for various sensing tasks. Albeit promising, since feature extraction and selection plays a key role in system performance, hand-crafted features are usually task-specific and not reusable for new tasks, hindering their usage for ubiquitous sensing.
Table 9.
Hand-crafted statistical pattern-based sensing.
| Year | Reference | Methodology | Performance | Base Signal | Sensing Range | Setting |
|---|---|---|---|---|---|---|
| 2020 | MultiSense [71] | ICA-based BSS | 0.73 bpm mean error | Constructed reference-CSI-based signal ratio | 4 m × 7.5 m | 200 Hz; single pair of Tx-Rx |
| 2020 | Wang et al. [45] | Statistical pattern analysis | 86% accuracy | PSD of CSI | 3.5 m | 10 Hz; single pair of Tx-Rx |
| 2021 | WiGesture [35] | MNP feature extraction | 92.8–94.5% accuracy | CSI ratio | 4 m × 7 m | 400 Hz; two pairs of Tx-Rx |
| 2021 | WiMonitor [48] | Doppler frequency and activity intensity pattern extraction | Not applicable | CSI ratio | Multi-room | 200 Hz; single pair of Tx-Rx |
| 2021 | WiPhone [74] | Ambient reflection-based pattern extraction | 0.31 bpm average error | CSI amplitude | Multi-room apartment | 50 Hz; single pair of Tx-Rx with LOS blocked |
| 2022 | HandGest [36] | Hand-centric feature extraction, i.e., DPV and MRV | 4.7 cm accuracy | CSI ratio | 1 m | 500 Hz; two pairs of Tx-Rx |
| 2022 | Niu et al. [64] | DFS-based velocity estimation + receiver selection | 96.05% accuracy | CSI ratio | 7 m × 9.8 m | 1000 Hz; six pairs of Tx-Rx |
| 2022 | Fan et al. [69] | 2D-antenna array-based signal parameter estimation | 6.9° median error for moving direction estimation; 16.6° mean error for rotation angle estimation | Time-reversal resonating strength of CSI | 28 m × 36.5 m | 200 Hz; single pair of Tx-Rx; half octagonal array of 6 antennas |
| 2022 | Wi-Drone [70] | Rigid-body coordinate transformation-based absolute pose and relative motion estimation | 26.1 cm average location accuracy and 3.8° rotation accuracy | CSI | 32 m × 18 m | Four pairs of Tx-Rx |
| 2022 | DPSense-WiGesture [37] | Signal segmentation + sensing quality-based signal processing | 94% average accuracy | CSI | 1.2 m | 400 Hz; two pairs of Tx-Rx |
| 2022 | Niu et al. [38] | Position-independent feature extraction, i.e., movement fragment and relative motion direction change | 96% accuracy | CSI ratio | 2 m × 2 m | 1000 Hz; 2 pairs of Tx-Rx |
| 2022 | WiCPD [23] | Feature-based motion, stationary and transition target detector | 96.56–100% real-time detection rate | ACF of CSI | Car | 30 Hz; single pair of Tx-Rx |
| 2023 | Hu et al. [24] | Sub-carrier correlation and covariance feature extraction | 95% and 99% true positive rate for distance-based and room-based detection | Power of CSI | Multi-room | 30 Hz; single pair of Tx-Rx |
| 2023 | WiTraj [65] | DFS extraction + multi-view trajectory estimation + motion detection | 2.5% median tracking error | CSI ratio | 7 m × 6 m | 400 Hz; three pairs of Tx-Rx |
| 2023 | Zhang et al. [67] | Intelligent reflecting surface construction | Sub-centimeter accuracy | Received signal power | 6 m × 6 m | Single pair of Tx-Rx |
| 2024 | Zhang et al. [68] | Intelligent reflecting surface construction | 0.11 m average error | Received signal power | 3 m × 3 m | Single pair of Tx-Rx |
| 2024 | Xie et al. [76] | Respiratory energy-based interference detection and convex optimization-based beam control | 32% mean absolute error reduction | CSI | 9 m × 6 m | Single pair of Tx-Rx |
| 2024 | WiCGesture [44] | Meta motion-based signal segmentation and back-tracking searching-based identification | 89.6% for digits and 88.3% for Greek letters | CSI ratio | 1 m | 400 Hz; Two pairs of Tx-Rx |
| 2024 | FewSense [66] | TD-CSI-based Doppler speed estimation | 34 cm median error | Time domain CSI difference | 7 m × 7 m | 1000 Hz; Two pairs of Tx-Rx |
| 2024 | WI-MOID [26] | Physical and statistical pattern extraction + SVM + state machine | 97.34% accuracy and 1.75% false alarm rate | ACF of CSI | Multi-room | 1500 Hz; single pair of Tx-Rx |
Automatic deep pattern extraction-based sensing. Since it is challenging to devise effective sensing features, more and more studies have begun leveraging various deep learning models for better accuracy and robustness, such as the Convolution Neural Network (CNN) and Recurrent Neural Network (RNN). As seen in Table 10, the combination of CNN and RNN has been widely adopted in recent works [27,30,32,39,73,77,79] due to its advantage in extracting spatial-temporal features from the CSI signal automatically. In addition, to gain more general representation learning, adversarial learning and few-shot learning have also been used for efficient and robust feature training [29,31,34,43,53,57,60,61]. The end-to-end nature of deep learning has made network framework selection and design the primary factor in sensing system implementations.
Table 10.
Automatic deep pattern-based sensing.
| Year | Reference | Methodology | Performance | Base Signal | Sensing Range | Setting |
|---|---|---|---|---|---|---|
| 2020 | WiPose [77] | CNN + LSTM | 2.83 cm average error | 3D velocity profile of CSI | Single room | 1000 Hz; three pairs of Tx-Rx; distributed deployed receiving Antennas |
| 2020 | WiSIA [78] | cGAN | Not applicable | Power of CSI | 2.1 m | 1000 Hz; two pairs od Tx-Rx; receiving antennas orthogonal to each other |
| 2021 | Kang et al. [34] | Adversarial learning and attention scheme | 3–12.7% improvement | DFS of CSI | 2 m × 2 m | Two pairs of Tx-Rx from Widar dataset |
| 2022 | GaitSense [27] | CNN + LSTM + transfer learning + data augmentation | 98% accuracy | Gait-BVP of CSI | 4.6 m × 4.4 m | 1000 Hz; six pairs of Tx-Rx |
| 2021 | Ma et al. [46] | CNN + reinforcement learning | 97% average accuracy | CSI amplitude | 6.8 m × 4 m | 100 Hz; single pair of Tx-Rx |
| 2021 | MCBAR [47] | GAN and semi-supervised learning | 90% average accuracy | CSI amplitude | 6.5 m × 6.3 m | single pair of Tx-Rx |
| 2021 | WiFi-Sleep [73] | Respiration and movement pattern extraction + CNN-BiLSTM | 81.8% accuracy | CSI ratio | Close to the bed | 200 Hz; single pair of Tx-Rx |
| 2022 | CAUTION [29] | Few-shot learning | 93.06 average accuracy | CSI amplitude | 5.2 m × 7.2 m | Single pair of Tx-Rx |
| 2022 | Ding et al. [50] | DCN + transfer learning | 96.85% average accuracy | CSI | 6 m × 8 m | 200 Hz; single pair of Tx-Rx |
| 2022 | EfficientFi [51] | DNN | 98% accuracy | CSI amplitude | 6.5 m × 5 m | 500 Hz; single pair of Tx-Rx |
| 2022 | GoPose [79] | 2D AOA spectrum + CNN + LSTM | 93.2% for 5 users and 76.2% for 11 users | CSI phase | 4 m × 4 m | 1000 Hz; four pairs of Tx-Rx; L-shaped receiving antennas |
| 2022 | ResFi [75] | CNN-based classification | 95% accuracy | CSI amplitude | 1 m | 10 Hz; single pair of Tx-Rx |
| 2022 | TOSS [52] | Meta learning + pseudo label strategy | 82.69% average accuracy | CSI | Single room | Single pair of Tx-Rx |
| 2022 | Widar 3.0 [39] | BVP feature + CNN-RNN | 92.7% in-domain and 82.6–92.4% cross-domain accuracy | BVP of CSI | 2 m × 2 m | 1000 Hz; six pairs of Tx-Rx |
| 2022 | WiFine [40] | data enhancement-based feature extraction + lightweight neural network | 96.03% accuracy in 0.19 s | CSI | Single room | Single pair of Tx-Rx |
| 2022 | Wi-PIGR [30] | Spectrogram optimization + CNN + LSTM | 93.5% for single user and 77.15% for 50 users | CSI amplitude | 5 m × 5 m | 1000 Hz; two pairs of Tx-Rx |
| 2023 | Auto-Fi [31] | Geometric self-supervised learning + few-shot calibration | 86.83% for gesture; 79.61% for gait | CSI amplitude | Single room | 100 Hz; single pair of Tx-Rx |
| 2023 | GaitFi [32] | RCN + LSTM + feature fusion | 94.2% accuracy | CSI + video | 2.1 m | 800 Hz; single pair of Tx-Rx |
| 2023 | MetaFi++ [81] | CNN + Transformer | 97.3% for PCK@50 | CSI + video | Single room | 1000 Hz; single pair of Tx-Rx |
| 2023 | FallDar [53] | Scattering model + VAE generative model + DNN adversarial learning model | 5.7% false alarm rate and 3.4% missed alarm rate | ACF of CSI | 3.6 m × 8.4 m | 1000 Hz; single pair of Tx-Rx |
| 2023 | SHARP [54] | Phase correction-based DFS extraction + Nerual network | 95% average accuracy | CSI | 5 m × 6 m | 173 Hz; single pair of Tx-Rx |
| 2023 | UniFi [41] | DFS extraction + consistency-guided multi-view deep network + mutual information-based regularization | 99% and 90–98% accuracy for in-domain and cross-domain recognition | CSI ratio | 2 m × 2 m | Widar dataset |
| 2023 | WiTransformer [42] | Transformer | 86.16% accuracy | BVP of CSI | 2 m × 2 m | Widar dataset |
| 2024 | AirFi [43] | Data augmentation + adversarial learning +domain generalization | 90% accuracy | CSI amplitude | 4 m × 4 m | Single pair of Tx-Rx |
| 2024 | i-Sample [57] | Intermediate sample generation + domain adversarial adaptation | 10% accuracy gain | CSI | Single room | Single pair of Tx-Rx |
| 2024 | MaskFi [58] | Transformer-based encoder + Gate Recurrent Unit network | 97.61% average accuracy | CSI + video | Single room | 1000 Hz; Single pair of Tx-Rx |
| 2024 | MetaFormer [59] | Transformer-based spatial-temporal feature extraction + match-based meta-learning approach | Improved accuracy in various cross-domain scenarios | CSI | Single room | SiFi, Widar, Wiar datasets |
| 2024 | PowerSkel [83] | Knowledge distillation network based on collaborative learning and self-attention | 96.27% for PCK@50 | CSI + Kinect video | Single room | Three pairs of Tx-Rx |
| 2024 | SAT [60] | Calibrated confidence-based adversarial sample selection + adversarial learning | Improved accuracy and robustness | CSI | Single room | Single pair of Tx-Rx |
| 2024 | SecureSense [61] | Consistency-guided adversarial learning | Robust performance under various attacks | CSI amplitude | 5 m × 6.5 m | 1000 Hz; single pair of Tx-Rx |
| 2024 | Luo et al. [62] | Transformer | 98.78% accuracy | CSI | Single room | UT-HAR dataset |
| 2024 | Wi-Diag [33] | Independent component analysis-based blind source separation + CycleGAN | 87.77% average accuracy | CSI | 7 m × 8 m | 1000 Hz; single pair of Tx-Rx |
| 2024 | WiSMLF [63] | High frequency energy-based sensing scheme selection + VGG/LSTM-based multi-level feature fusion | 92% average accuracy | CSI | Single room | 100 Hz; single pair of Tx-Rx |
| 2024 | Zhu et al. [25] | ResNet18 | 95.57% average accuracy | Amplified ACF of CSI | 6 m × 6.5 m | 1500 Hz; single pair of Tx-Rx |
Apart from the above differences, we can obtain several additional findings from Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10. First, apart from the CSI amplitude and phase information, several new base signals, such as the BVP of CSI, ACF of CSI, and CSI ratio, have been used for alleviating the intrinsic errors of COTS WiFi devices [88]. Among these base signals, the CSI ratio is drawing more attention since it can not only remove the CSI offset, but it can also increase the sensing signal-to-noise rate (SNR) [89]. Second, some works have tried to combine a pattern-based scheme with model-based scheme to ensure the performance and reliability of complex sensing applications. Third, many systems have been developed for single human sensing under constrained deployment, i.e., single room sensing area with the LOS condition satisfied.
4. Challenges
Despite the above endeavors devoted to bringing Wi-Fi sensing from laboratory study to real-life applications, either by improving sensing granularity or exploring application scenarios, most of the existing works still face great practical challenges. Specifically, making Wi-Fi sensing system readily available for wide real-world deployment, easily adaptable to different environments, and with enough sensing coverage is of vital importance. This section presents two key challenges faced in existing works, i.e., the domain dependent issue and the sensing range limitation, and it discusses related potential solution explorations.
Domain dependent issue. As the superposition result of multi-path signals, Wi-Fi is highly sensitive to various factors, such as locations, orientations, targets, and environments. This is also known as the domain-dependence problem [15,18,86]. For example, the same human activity will lead to quite different CSI variations if the location or orientation of the target changes, as revealed by the Fresnel zone model. Moreover, different sensing environments and device settings will make this inconsistent phenomenon even worse. A sensing system lacking resilience to domain variations is in fact of little practical use for ubiquitous sensing. Thus, in order to make Wi-Fi sensing reusable and robust among different settings, researchers have explored various methods, as summarized in Table 11. Since training effort accounts for a great part of the system deployment cost, Table 11 classifies the related works into three categories, i.e., training-free, training-once, and training + Calibration/Retrained. As seen in Table 11, the training-free scheme is mainly used for simple presence detection tasks [23,24], where a motion-induced threshold is predetermined without training. In addition, for the training-once scheme, the domain-independent feature extraction is the most studied, and it is mainly used along model-based sensing, as listed in Table 8 due to clear interpretability. Moreover, with the increasing complexity of sensing tasks and environments, system recalibration would become inevitable, promoting researchers to reduce the system retraining cost, e.g., utilizing data augmentation, transfer learning, and few-shot learning, as shown in Table 11. It can also be observed that domain-independent feature extraction can be used alone or further integrated with other retraining algorithms. Drawn from the above discussions, for these complex applications, it is expected that combining the strengths of model-based and auto deep learning model-based methods can enable a more general and robust Wi-Fi sensing realization.
Table 11.
Cross-domain Wi-Fi sensing.
| Training Cost | Cross-Domain Scheme | Related Work |
|---|---|---|
| Training-free | Domain-independent feature extraction | [23,24] |
| Training-once | Domain-independent feature extraction | [25,26,27,28,30,34,35,36,37,38,39,41,42,44,49,53,54,64,65,66,72] |
| Training + Calibration/Retrained | Generative adversarial network | [33,47,53,61] |
| Transfer learning | [27,31,34,43,50,57,60] | |
| Few-shot learning | [29,31,43,52] | |
| Data augmentation | [27,43,57] | |
| CNN +LSTM/GRU/Transformer | [25,30,32,39,41,42,46,58,59,62,81] |
Sensing range limitation. As illustrated in the tables of last section, the existing sensing range is usually just 6–8 m within a single room, while the communication range of Wi-Fi can reach tens of meters, greatly hindering real-world applications. The short sensing range is mainly because Wi-Fi sensing relies on target-induced reflection signal variation, which is much weaker compared to direct LOS signal and contains intrinsic hardware noise. To be more specific, due to hardware imperfections and clock synchronization errors, the raw CSI amplitude contains high impulse and burst noise, while the raw randomly corrupted CSI phase is even more unusable in practice. To deal with this limitation, some researchers proposed employing a new base signal derived from the raw CSI, namely the CSI ratio as seen in Table 8 and Table 9. Defined as the quotient of CSI readings between two receiver antennas, the CSI ratio can remove the amplitude noise and phase noise effectively. More specifically, since different antennas on the same receiver share the same RF chain and clock, the division operation can cancel out most of the noise, gaining a more ideal amplitude and a phase signal with a high signal-to-noise ratio (SNR). The higher SNR and phase usability of the CSI ratio serve as the key enablers for the longer sensing range and higher sensing accuracy. FarSense [90] first increased fine-grained sensing range to 8 m using the CSI ratio signal, while Zeng et al. [91] and DiverSense [92] further boosted the sensing range to 18 m and 40 m by fully utilizing the spatial and frequency diversity. In addition to constructing a new base signal, Wang et al. [93] studied the effect of device placement on sensing SNR and doubly expanded the sensing range by properly placing the transmitter and receiver. Overall, sensing range enlargement is pivoted to large-scale sensing applications and is still in its infancy, requiring further exploration and validation in complex real-world scenario deployments.
5. Future Research Trend Discussion
Despite the great effort spent on Wi-Fi sensing over the past years, there still exists a great gap for its pervasive real-life application. Based on the detailed analysis above, we point out three critical barriers that require further research in this section.
Table 12.
CSI extraction tools.
| Year | CSI Extraction Tool | IEEE Standard | Related Work |
|---|---|---|---|
| 2011 | 802.11n CSI Tool [17] | 802.11n | [27,28,30,33,35,36,37,38,39,44,46,48,49,50,52,53,56,57,60,63,64,65,66,71,73,77,78,79,80,82,84] |
| 2015 | Atheros CSI Tool [94] | 802.11n | [29,31,32,47,51,58,61,72,81,94] |
| 2019 | Nexmon CSI [95] | 802.11 ac | [40,54,74,75,95] |
| 2020 | ESP32 CSI Tool [96,97] | Any computer, smartphone or even standalone | [83,96,97] |
| 2021 | AX-CSI [98] | 802.11 ax | [98] |
| 2022 | PicoScenes [99] | 802.11 a/g/n/ac/ax | [70,99] |
Table 13.
Wi-Fi sensing datasets.
| Year | Dataset | Description | Tool | Related Work |
|---|---|---|---|---|
| 2017 | UT-HAR [100] | Activity data | 802.11n CSI Tool | [31,46,62] |
| 2018 | SignFi [101] | Sign data | 802.11n CSI Tool | [40,59] |
| 2018 | FallDeFi [102] | Fall data | 802.11n CSI Tool | [46,53] |
| 2019 | WiAR [103] | Activity and gesture data | 802.11n CSI Tool | [59] |
| 2019 | Widar [104] | Gesture data | 802.11n CSI Tool | [31,34,39,41,42,43,59] |
| 2021 | OneFi [105] | Gesture data | 802.11n CSI Tool | [105] |
| 2023 | MM-Fi [106] | Multi-modal dataset | Atheros CSI Tool | [58] |
| 2023 | NTU-Fi [107] | Activity and gait data | Atheros CSI Tool | [62] |
| 2023 | SHARP [54] | Activity data | Nexmon CSI | [54] |
| 2023 | Cominelli [108] | Activity data | AX-CSI | [108] |
| 2023 | WiTraj [65] | Trajectory data | 802.11n CSI Tool | [65] |
Sensing assessment standardization. One key issue is the lack of a standard performance evaluation of the various Wi-Fi sensing systems. Unlike the widely accepted standard evaluation criterion in the computer vision domain, there is still a lack of an effective and consistent testing platform in Wi-Fi sensing. Specifically, the deficiency exists in two aspects, i.e., CSI extraction tool diversity and evaluation dataset scarcity. The diversity of CSI extraction tools is shown in Table 12, with Intel 5300 NIC-based 802.11n CSI Tool being the most popular one used. However, sensing techniques developed with old 802.11n protocol have not explored the innovations of newer standards and may even fail when used on new-generation Wi-Fi cards [108,109]. In addition, as illustrated in Table 13, although there have been some publicly released datasets, none of them have been widely used. Existing works mostly adopt self-collected datasets, collected in different scenarios with different tools, hindering the comparability and replicability of research outcomes. To build comprehensive datasets without labor-intensive and time-consuming efforts, researchers have studied radio signal synthesis [110,111] and physical data augmentation [112], providing promising solutions to the data scarcity problem. We believe a more unified CSI extraction tool compatible with the new 802.11 standard and a set of standard datasets for a benchmark comparison should be indispensable for the further research cooperation and development of Wi-Fi sensing.
Sensing and communication balance. As illustrated in Table 14, most sensing systems require a high sampling rate for reliable performance, which interferes with regular Wi-Fi communication. To be more specific, the data throughput undergoes great drop when the sampling rate for sensing is higher than 50 Hz [66]. SenCom [113] managed to extract CSI from general communication packets and obtained evenly sampled and sufficient CSI data with a detailed signal processing technique. While appealing, SenCom is not yet applicable for COTS clients. Thus, the ways of enabling Wi-Fi sensing while maintaining communication capability, i.e., achieving sensing and communication balance, remain an open problem in the current ISAC area.
Sensing generalization and reliability. As noted in Table 12, raw CSI reading is still only accessible with limited hardware; some researchers resorted to sensing with other Wi-Fi signals. For instance, since the beamforming feedback matrix (BFM) is readily available with all new-generation MU-MIMO-enabled Wi-Fi cards, researchers have explored generalized Wi-Fi sensing using BFM [114,115]. In addition, to improve the reliability of sensing, multi-modal sensing, which integrates Wi-Fi and other sensing modalities, e.g., video [32,52,81,116] and received signal strength indicator (RSSI) [117], are worth further studying.
Table 14.
Sampling rate of recent works.
Apart from the above discussion, the physical challenges of the existing Wi-Fi infrastructure should also be noticed, which will greatly determine the possible sensing limit of Wi-Fi sensing. First, due to hardware and network design, clock asynchronism between Wi-Fi transmitter and receiver is a severe issue in an ISAC system. It introduces a time-varying random phase offset in raw CSI, making reliable feature extraction difficult. Second, except for target influence, dynamic parameter adjustments of the network card during transmission also affect the CSI measurement, which is highly dependent on the hardware design. Third, large-scale Wi-Fi sensing needs to obtain CSI from multiple distributed receivers. The ways of enabling CSI estimation and alignment over multiple devices are a challenging problem. Currently, there is no universal solution to the above challenges, requiring cooperative efforts from application researchers, chip manufactures, and communication equipment vendors.
6. Conclusions
Owing to the active participation from numerous researchers, notable advances have been made in Wi-Fi sensing techniques in recent years. In an effort to gain insight into future trends, this paper reviews major achievements over the last 5 years and carries out an in-depth analysis of various methods, including limitations and practical challenges faced in existing systems. Moreover, to realize massive real-life applications, this paper highlights three imperative and promising future directions which are as follows: sensing assessment standardization, sensing and communication balance, and sensing generalization and reliability. We hope this review can help people better understand the progress and problems within the current Wi-Fi sensing research field, inspiring more amazing ideas for the upcoming ubiquitous ISAC.
Appendix A
Table A1 summaries all references mentioned in Section 3, pointing out corresponding application and methodology categories, and can direct the interested reader to related subsections for a more detailed description.
Table A1.
Summary of references in Section 3.
| Year | Reference | Application | Methodology | Related Subsections |
|---|---|---|---|---|
| 2022 | WiCPD [23] | In-car child presence detection | Feature-based motion, stationary and transition target detector |
Table 1 in Section 3.1; Table 9 in Section 3.2 |
| 2023 | Hu et al. [24] | Proximity detection | Sub-carrier correlation and covariance feature extraction |
Table 1 in Section 3.1; Table 9 in Section 3.2 |
| 2024 | Zhu et al. [25] | Human and non-human differentiation | ResNet18 |
Table 1 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | WI-MOID [26] | Edge device-based human and non-human differentiation | Physical and statistical pattern extraction + SVM + state machine |
Table 1 in Section 3.1; Table 9 in Section 3.2 |
| 2021 | GaitSense [27] | Gait-based human identification | CNN + LSTM + transfer learning + data augmentation |
Table 2 in Section 3.1; Table 10 in Section 3.2 |
| 2021 | GaitWay [28] | Gait speed estimation | Scattering model |
Table 2 in Section 3.1; Table 8 in Section 3.2 |
| 2022 | CAUTION [29] | Gait-based human authentication | Few-shot learning |
Table 2 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | Wi-PIGR [30] | Gait recognition | Spectrogram optimization + CNN + LSTM |
Table 2 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | Auto-Fi [31] | Gesture and gait recognition | Geometric self-supervised learning + few-shot calibration |
Table 2 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | GaitFi [32] | Gait recognition | RCN + LSTM + feature fusion |
Table 2 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | Wi-Diag [33] | Multi-subject abnormal gait diagnosis | Independent component analysis-based blind source separation + CycleGAN |
Table 2 in Section 3.1; Table 10 in Section 3.2 |
| 2021 | Kang et al. [34] | Gesture recognition | Adversarial learning and attention scheme |
Table 3 in Section 3.1; Table 10 in Section 3.2 |
| 2021 | WiGesture [35] | Gesture recognition | MNP feature extraction |
Table 3 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | HandGest [36] | Handwriting recognition | Hand-centric feature extraction, i.e., DPV and MRV |
Table 3 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | DPSense-WiGesture [37] | Gesture recognition | Signal segmentation + sensing quality-based signal processing |
Table 3 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | Niu et al. [38] | Gesture recognition | Position-independent feature extraction, i.e., movement fragment and relative motion direction change |
Table 3 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | Widar 3.0 [39] | Cross-domain gesture recognition | BVP feature + CNN-RNN |
Table 3 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | WiFine [40] | Gesture recognition | Data enhancement-based feature extraction + lightweight neural network |
Table 3 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | UniFi [41] | Gesture recognition | DFS extraction + consistency-guided multi-view deep network + mutual information-based regularization |
Table 3 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | WiTransformer [42] | Gesture recognition | Transformer |
Table 3 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | AirFi [43] | Gesture recognition | Data augmentation + adversarial learning + domain generalization |
Table 3 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | WiCGesture [44] | Continuous gesture recognition | Meta motion-based signal segmentation and back-tracking searching-based identification |
Table 3 in Section 3.1; Table 9 in Section 3.2 |
| 2020 | Wang et al. [45] | People counting and recognition | Statistical pattern analysis |
Table 4 in Section 3.1; Table 9 in Section 3.2 |
| 2021 | Ma et al. [46] | Activity recognition | CNN + reinforcement learning |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2021 | MCBAR [47] | Activity recognition | GAN and semi-supervised learning |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2021 | WiMonitor [48] | Location and activity monitoring | Doppler frequency and activity intensity pattern extraction |
Table 4 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | DeFall [49] | Fall detection | Scattering model |
Table 4 in Section 3.1; Table 8 in Section 3.2 |
| 2022 | Ding et al. [50] | Activity recognition | DCN + transfer learning |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | EfficientFi [51] | Activity recognition | DNN |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | TOSS [52] | Activity recognition | Meta learning + pseudo label strategy |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | FallDar [53] | Fall detection | Scattering model + VAE generative model + DNN adversarial learning model |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | SHARP [54] | Activity recognition | Phase correction-based DFS extraction + Nerual network |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | Liu et al. [55] | Moving receiver-based activity recognition | Dynamic Fresnel zone model |
Table 4 in Section 3.1; Table 8 in Section 3.2 |
| 2023 | WiCross [56] | Target passing detection | Diffraction model-based phase pattern extraction |
Table 4 in Section 3.1; Table 8 in Section 3.2 |
| 2024 | i-Sample [57] | Activity recognition | Intermediate sample generation + domain adversarial adaptation |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | MaskFi [58] | Activity recognition | Transformer-based encoder + Gate Recurrent Unit network |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | MetaFormer [59] | Activity recognition | Transformer-based spatial-temporal feature extraction + match-based meta-learning approach |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | SAT [60] | Activity recognition | Calibrated confidence-based adversarial sample selection + adversarial learning |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | SecureSense [61] | Activity recognition under adversarial attack | Consistency-guided adversarial learning |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | Luo et al. [62] | Activity recognition | Transformer |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | WiSMLF [63] | Activity recognition | High-frequency energy-based sensing scheme selection + VGG/LSTM-based multi-level feature fusion |
Table 4 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | Niu et al. [64] | Velocity estimation-based tracing | DFS-based velocity estimation + receiver selection |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2023 | WiTraj [65] | Human walking tracking | DFS extraction + multi-view trajectory estimation + motion detection |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2024 | FewSense [66] | Tracking | TD-CSI-based Doppler speed estimation |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2023 | Zhang et al. [67] | Multi-person localization | 2D antenna array-based signal parameter estimation |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2024 | Zhang et al. [68] | Passive localization | Rigid-body coordinate transformation-based absolute pose and relative motion estimation |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | Fan et al. [69] | Moving direction estimation | Intelligent reflecting surface construction |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | Wi-Drone [70] | Tracking-based indoor drone flight control | Intelligent reflecting surface construction |
Table 5 in Section 3.1; Table 9 in Section 3.2 |
| 2020 | MultiSense [71] | Multi-person respiration sensing | ICA-based BSS |
Table 6 in Section 3.1; Table 9 in Section 3.2 |
| 2021 | SMARS [72] | Breath estimation and sleep stage recognition | Scattering model |
Table 6 in Section 3.1; Table 8 in Section 3.2; |
| 2021 | WiFi-Sleep [73] | Sleep stage monitoring | Respiration and movement pattern extraction + CNN-BiLSTM |
Table 6 in Section 3.1; Table 10 in Section 3.2 |
| 2021 | WiPhone [74] | Respiration monitoring | Ambient reflection-based pattern extraction |
Table 6 in Section 3.1; Table 9 in Section 3.2 |
| 2022 | ResFi [75] | Respiration detection | CNN-based classification |
Table 6 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | Xie et al. [76] | Respiration sensing with interfering individual | Respiratory energy-based interference detection and convex optimization-based beam control |
Table 6 in Section 3.1; Table 9 in Section 3.2 |
| 2020 | WiPose [77] | Pose construction | CNN + LSTM |
Table 7 in Section 3.1; Table 10 in Section 3.2 |
| 2020 | WiSIA [78] | Target imaging | cGAN |
Table 7 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | GoPose [79] | 3D human pose estimation | 2D AOA spectrum + CNN + LSTM |
Table 7 in Section 3.1; Table 10 in Section 3.2 |
| 2022 | Wiffract [80] | Still object imaging | Keller’s Geometrical Theory of Diffraction |
Table 7 in Section 3.1; Table 8 in Section 3.2 |
| 2023 | MetaFi++ [81] | Pose estimation | CNN + Transformer |
Table 7 in Section 3.1; Table 10 in Section 3.2 |
| 2023 | WiMeasure [82] | Object size measurement | Diffraction model |
Table 7 in Section 3.1; Table 8 in Section 3.2 |
| 2024 | PowerSkel [83] | Pose estimation | Knowledge distillation network based on collaborative learning and self-attention |
Table 7 in Section 3.1; Table 10 in Section 3.2 |
| 2024 | WiProfile [84] | 2D target Profiling | Diffraction effect-based profiling + inverse Fresnel transform |
Table 7 in Section 3.1; Table 8 in Section 3.2 |
Author Contributions
Conceptualization, H.Z.; writing—original draft preparation, H.Z., E.D. and M.X.; discussion and supervision, H.L. and F.W. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This work was supported in part by the Young Scientists Fund of the National Natural Science Foundation of China, under grants 61902237 and 52205597, and the Key Project of Science and Technology Commission of Shanghai Municipality, under grant 22DZ1100803.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Liu A., Huang Z., Li M., Wan Y., Li W., Han T.X., Liu C., Du R., Tan D.K.P., Lu J., et al. A Survey on Fundamental Limits of Integrated Sensing and Communication. IEEE Commun. Surv. Tutor. 2022;24:994–1034. doi: 10.1109/COMST.2022.3149272. [DOI] [Google Scholar]
- 2.Liu F., Cui Y., Masouros C., Xu J., Han T.X., Eldar Y.C., Buzzi S. Integrated Sensing and Communications: Toward Dual-Functional Wireless Networks for 6G and Beyond. IEEE J. Sel. Areas Commun. 2022;40:1728–1767. doi: 10.1109/JSAC.2022.3156632. [DOI] [Google Scholar]
- 3.Meneghello F., Chen C., Cordeiro C., Restuccia F. Toward Integrated Sensing and Communications in IEEE 802.11bf Wi-Fi Networks. IEEE Commun. Mag. 2023;61:128–133. doi: 10.1109/MCOM.001.2200806. [DOI] [Google Scholar]
- 4.Wu C., Wang B., Au O., Liu K. Wi-Fi Can Do More: Toward Ubiquitous Wireless Sensing. IEEE Commun. Stand. Mag. 2022;6:42–49. doi: 10.1109/MCOMSTD.0001.2100111. [DOI] [Google Scholar]
- 5.Li X., Cui Y., Zhang J., Liu F., Zhang D., Hanzo L. Integrated Human Activity Sensing and Communications. IEEE Commun. Mag. 2023;61:90–96. doi: 10.1109/MCOM.002.2200391. [DOI] [Google Scholar]
- 6.Yang Z., Zhou Z., Liu Y. From RSSI to CSI: Indoor localization via channel response. ACM Comput. Surv. 2013;46:1–32. doi: 10.1145/2543581.2543592. [DOI] [Google Scholar]
- 7.Zhang F., Wu C., Wang B., Lai H., Han Y., Ray Liu K. WiDetect: Robust Motion Detection with a Statistical Electromagnetic Model. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019;3:1–24. doi: 10.1145/3351280. [DOI] [Google Scholar]
- 8.Zhang F., Niu K., Xiong J., Jin B., Gu T., Jiang Y., Zhang D. Towards a Diffraction-based Sensing Approach on Human Activity Recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019;3:1–25. doi: 10.1145/3314420. [DOI] [Google Scholar]
- 9.Gong W., Liu J. SiFi: Pushing the Limit of Time-Based WiFi Localization Using a Single Commodity Access Point. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;2:1–21. doi: 10.1145/3191742. [DOI] [Google Scholar]
- 10.Zhang D., Wang H., Wu D. Toward Centimeter-Scale Human Activity Sensing with Wi-Fi Signals. Computer. 2017;50:48–57. doi: 10.1109/MC.2017.7. [DOI] [Google Scholar]
- 11.Wang Z., Jiang K., Hou Y., Dou W., Zhang C., Huang Z., Guo Y. A Survey on Human Behavior Recognition Using Channel State Information. IEEE Access. 2019;7:155986–156024. doi: 10.1109/ACCESS.2019.2949123. [DOI] [Google Scholar]
- 12.Ma Y., Zhou G., Wang S. WiFi Sensing with Channel State Information: A Survey. ACM Comput. Surv. 2019;52:1–36. doi: 10.1145/3310194. [DOI] [Google Scholar]
- 13.Tan S., Ren Y., Yang J., Chen Y. Commodity WiFi Sensing in Ten Years: Status, Challenges, and Opportunities. IEEE Internet Things J. 2022;9:17832–17843. doi: 10.1109/JIOT.2022.3164569. [DOI] [Google Scholar]
- 14.Xiao J., Li H., Wu M., Jin H., Jamal Deen M., Cao J. A Survey on Wireless Device-free Human Sensing: Application Scenarios, Current Solutions, and Open Issues. ACM Comput. Surv. 2022;55:1–35. doi: 10.1145/3530682. [DOI] [Google Scholar]
- 15.Chen C., Zhou G., Lin Y. Cross-Domain WiFi Sensing with Channel State Information: A Survey. ACM Comput. Surv. 2023;55:1–37. doi: 10.1145/3570325. [DOI] [Google Scholar]
- 16.Halperin D., Hu W., Sheth A., Wetherall D. Predictable 802.11 packet delivery from wireless channel measurements. SIGCOMM Comput. Commun. Rev. 2010;40:159–170. doi: 10.1145/1851275.1851203. [DOI] [Google Scholar]
- 17.Halperin D., Hu W., Sheth A., Wetherall D. Tool release: Gathering 802.11n traces with channel state information. SIGCOMM Comput. Commun. Rev. 2011;41:53. doi: 10.1145/1925861.1925870. [DOI] [Google Scholar]
- 18.Wang H., Zhang D., Ma J., Wang Y., Wang Y., Wu D., Gu T., Xie B. Human respiration detection with commodity wifi devices: Do user location and body orientation matter?; Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp ‘16); Heidelberg, Germany. 12–16 September 2016; New York, NY, USA: Association for Computing Machinery; 2016. pp. 25–36. [Google Scholar]
- 19.Wu D., Zhang D., Xu C., Wang H., Li X. Device-Free WiFi Human Sensing: From Pattern-Based to Model-Based Approaches. IEEE Commun. Mag. 2017;55:91–97. doi: 10.1109/MCOM.2017.1700143. [DOI] [Google Scholar]
- 20.Zhang F., Zhang D., Xiong J., Wang H., Niu K., Jin B., Wang Y. From Fresnel Diffraction Model to Fine-grained Human Respiration Sensing with Commodity Wi-Fi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;2:1–23. doi: 10.1145/3191785. [DOI] [Google Scholar]
- 21.Yang Z., Zhang Y., Chi G., Zhang G. Hands-on wireless sensing with Wi-Fi: A tutorial. arXiv. 20222206.09532 [Google Scholar]
- 22.Zhang F., Chen C., Wang B., Liu K.J.R. WiSpeed: A Statistical Electromagnetic Approach for Device-Free Indoor Speed Estimation. IEEE Internet Things J. 2018;5:2163–2177. doi: 10.1109/JIOT.2018.2826227. [DOI] [Google Scholar]
- 23.Zeng X., Wang B., Wu C., Regani S.D., Liu K.J.R. WiCPD: Wireless Child Presence Detection System for Smart Cars. IEEE Internet Things J. 2022;9:24866–24881. doi: 10.1109/JIOT.2022.3194873. [DOI] [Google Scholar]
- 24.Hu Y., Ozturk M.Z., Wang B., Wu C., Zhang F., Liu K.J.R. Robust Passive Proximity Detection Using Wi-Fi. IEEE Internet Things J. 2023;10:6221–6234. doi: 10.1109/JIOT.2022.3224701. [DOI] [Google Scholar]
- 25.Zhu G., Wang B., Gao W., Hu Y., Wu C., Liu K.J.R. WiFi-Based Robust Human and Non-human Motion Recognition With Deep Learning; Proceedings of the 2024 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops); Biarritz, France. 11–15 March 2024; pp. 769–774. [Google Scholar]
- 26.Zhu G., Hu Y., Wang B., Wu C., Zeng X., Liu K.J.R. Wi-MoID: Human and Nonhuman Motion Discrimination Using WiFi With Edge Computing. IEEE Internet Things J. 2024;11:13900–13912. doi: 10.1109/JIOT.2023.3339544. [DOI] [Google Scholar]
- 27.Zhang Y., Zheng Y., Zhang G., Qian K., Qian C., Yang Z. GaitSense: Towards Ubiquitous Gait-Based Human Identification with Wi-Fi. ACM Trans. Sens. Netw. 2021;18:1–24. doi: 10.1145/3466638. [DOI] [Google Scholar]
- 28.Wu C., Zhang F., Hu Y., Liu K.J.R. GaitWay: Monitoring and Recognizing Gait Speed Through the Walls. IEEE Trans. Mob. Comput. 2021;20:2186–2199. doi: 10.1109/TMC.2020.2975158. [DOI] [Google Scholar]
- 29.Wang D., Yang J., Cui W., Xie L., Sun S. CAUTION: A Robust WiFi-Based Human Authentication System via Few-Shot Open-Set Recognition. IEEE Internet Things J. 2022;9:17323–17333. doi: 10.1109/JIOT.2022.3156099. [DOI] [Google Scholar]
- 30.Zhang L., Wang C., Zhang D. Wi-PIGR: Path Independent Gait Recognition With Commodity Wi-Fi. IEEE Trans. Mob. Comput. 2022;21:3414–3427. doi: 10.1109/TMC.2021.3052314. [DOI] [Google Scholar]
- 31.Yang J., Chen X., Zou H., Wang D., Xie L. AutoFi: Toward Automatic Wi-Fi Human Sensing via Geometric Self-Supervised Learning. IEEE Internet Things J. 2023;10:7416–7425. doi: 10.1109/JIOT.2022.3228820. [DOI] [Google Scholar]
- 32.Deng L., Yang J., Yuan S., Zou H., Lu C.X., Xie L. GaitFi: Robust Device-Free Human Identification via WiFi and Vision Multimodal Learning. IEEE Internet Things J. 2023;10:625–636. doi: 10.1109/JIOT.2022.3203559. [DOI] [Google Scholar]
- 33.Zhang L., Ma Y., Fan X., Fan X., Zhang Y., Chen Z., Chen X., Zhang D. Wi-Diag: Robust Multisubject Abnormal Gait Diagnosis With Commodity Wi-Fi. IEEE Internet Things J. 2024;11:4362–4376. doi: 10.1109/JIOT.2023.3301908. [DOI] [Google Scholar]
- 34.Kang H., Zhang Q., Huang Q. Context-Aware Wireless-Based Cross-Domain Gesture Recognition. IEEE Internet Things J. 2021;8:13503–13515. doi: 10.1109/JIOT.2021.3064890. [DOI] [Google Scholar]
- 35.Gao R., Zhang M., Zhang J., Li Y., Yi E., Wu D., Wang L., Zhang D. Towards Position-Independent Sensing for Gesture Recognition with Wi-Fi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021;5:1–28. doi: 10.1145/3463504. [DOI] [Google Scholar]
- 36.Zhang J., Li Y., Xiong H., Dou D., Miao C., Zhang D. HandGest: Hierarchical Sensing for Robust-in-the-Air Handwriting Recognition With Commodity WiFi Devices. IEEE Internet Things J. 2022;9:19529–19544. doi: 10.1109/JIOT.2022.3170157. [DOI] [Google Scholar]
- 37.Gao R., Li W., Xie Y., Yi E., Wang L., Wu D., Zhang D. Towards Robust Gesture Recognition by Characterizing the Sensing Quality of WiFi Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–26. doi: 10.1145/3517241. [DOI] [Google Scholar]
- 38.Niu K., Zhang F., Wang X., Lv Q., Luo H., Zhang D. Understanding WiFi Signal Frequency Features for Position-Independent Gesture Sensing. IEEE Trans. Mob. Comput. 2022;21:4156–4171. doi: 10.1109/TMC.2021.3063135. [DOI] [Google Scholar]
- 39.Zheng Y., Zheng Y., Qian K., Zhang G., Liu Y., Wu C., Yang Z. Widar3.0: Zero-Effort Cross-Domain Gesture Recognition with Wi-Fi. IEEE Trans. Pattern Anal. Mach. Intell. 2022;44:8671–8688. doi: 10.1109/TPAMI.2021.3105387. [DOI] [PubMed] [Google Scholar]
- 40.Xing T., Yang Q., Jiang Z., Fu X., Wang J., Wu C.Q., Chen X. WiFine: Real-Time Gesture Recognition Using Wi-Fi with Edge Intelligence. ACM Trans. Sens. Netw. 2022;19:1–24. doi: 10.1145/3532094. [DOI] [Google Scholar]
- 41.Liu Y., Yu A., Wang L., Guo B., Li Y., Yi E., Zhang D. UniFi: A Unified Framework for Generalizable Gesture Recognition with Wi-Fi Signals Using Consistency-guided Multi-View Networks. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;7:1–29. doi: 10.1145/3631429. [DOI] [Google Scholar]
- 42.Yang M., Zhu H., Zhu R., Wu F., Yin L., Yang Y. WiTransformer: A Novel Robust Gesture Recognition Sensing Model with WiFi. Sensors. 2023;23:2612. doi: 10.3390/s23052612. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Wang D., Yang J., Cui W., Xie L., Sun S. AirFi: Empowering WiFi-Based Passive Human Gesture Recognition to Unseen Environment via Domain Generalization. IEEE Trans. Mob. Comput. 2024;23:1156–1168. doi: 10.1109/TMC.2022.3230665. [DOI] [Google Scholar]
- 44.Gao R., Li W., Liu J., Dai S., Zhang M., Wang L., Zhang D. WiCGesture: Meta-Motion-Based Continuous Gesture Recognition With Wi-Fi. IEEE Internet Things J. 2024;11:15087–15099. doi: 10.1109/JIOT.2023.3343875. [DOI] [Google Scholar]
- 45.Wang F., Zhang F., Wu C., Wang B., Liu K.J.R. Respiration Tracking for People Counting and Recognition. IEEE Internet Things J. 2020;7:5233–5245. doi: 10.1109/JIOT.2020.2977254. [DOI] [Google Scholar]
- 46.Ma Y., Arshad S., Muniraju S., Torkildson E., Rantala E., Doppler K., Zhou G. Location- and Person-Independent Activity Recognition with WiFi, Deep Neural Networks, and Reinforcement Learning. ACM Trans. Internet Things. 2021;2:1–25. doi: 10.1145/3424739. [DOI] [Google Scholar]
- 47.Wang D., Yang J., Cui W., Xie L., Sun S. Multimodal CSI-Based Human Activity Recognition Using GANs. IEEE Internet Things J. 2021;8:17345–17355. doi: 10.1109/JIOT.2021.3080401. [DOI] [Google Scholar]
- 48.Niu X., Li S., Zhang Y., Liu Z., Wu D., Shah R.C., Tanriover C., Lu H., Zhang D. WiMonitor: Continuous Long-Term Human Vitality Monitoring Using Commodity Wi-Fi Devices. Sensors. 2021;21:751. doi: 10.3390/s21030751. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Hu Y., Zhang F., Wu C., Wang B., Liu K.J.R. DeFall: Environment-Independent Passive Fall Detection Using WiFi. IEEE Internet Things J. 2022;9:8515–8530. doi: 10.1109/JIOT.2021.3116136. [DOI] [Google Scholar]
- 50.Ding X., Hu C., Xie W., Zhong Y., Yang J., Jiang T. Device-Free Multi-Location Human Activity Recognition Using Deep Complex Network. Sensors. 2022;22:6178. doi: 10.3390/s22166178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Yang J., Chen X., Zou H., Wang D., Xu Q., Xie L. EfficientFi: Toward Large-Scale Lightweight WiFi Sensing via CSI Compression. IEEE Internet Things J. 2022;9:13086–13095. doi: 10.1109/JIOT.2021.3139958. [DOI] [Google Scholar]
- 52.Zhou Z., Wang F., Yu J., Ren J., Wang Z., Gong W. Target-oriented Semi-supervised Domain Adaptation for WiFi-based HAR; Proceedings of the IEEE INFOCOM 2022-IEEE Conference on Computer Communications; London, UK. 2–5 May 2022; pp. 420–429. [Google Scholar]
- 53.Yang Z., Zhang Y., Zhang Q. Rethinking Fall Detection With Wi-Fi. IEEE Trans. Mob. Comput. 2023;22:6126–6143. doi: 10.1109/TMC.2022.3188779. [DOI] [Google Scholar]
- 54.Meneghello F., Garlisi D., Di Fabbro N., Tinnirello I., Rossi M. SHARP: Environment and Person Independent Activity Recognition With Commodity IEEE 802.11 Access Points. IEEE Trans. Mob. Comput. 2023;22:6160–6175. doi: 10.1109/TMC.2022.3185681. [DOI] [Google Scholar]
- 55.Liu J., Li W., Gu T., Gao R., Chen B., Zhang F., Wu D., Zhang D. Towards a Dynamic Fresnel Zone Model to WiFi-based Human Activity Recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2023;7:1–24. doi: 10.1145/3596270. [DOI] [Google Scholar]
- 56.Shi W., Wang X., Niu K., Wang L., Zhang D. WiCross: I Can Know When You Cross Using COTS WiFi Devices; Proceedings of the Adjunct Proceedings of the 2023 ACM International Joint Conference on Pervasive and Ubiquitous Computing & the 2023 ACM International Symposium on Wearable Computing (UbiComp/ISWC ‘23 Adjunct); Cancun, Mexico. 8–12 October 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 133–136. [Google Scholar]
- 57.Zhou Z., Wang F., Gong W. I-Sample: Augment Domain Adversarial Adaptation Models for WiFi-based HAR. ACM Trans. Sens. Netw. 2024;20:1–20. doi: 10.1145/3616494. [DOI] [Google Scholar]
- 58.Yang J., Tang S., Xu Y., Zhou Y., Xie L. MaskFi: Unsupervised Learning of WiFi and Vision Representations for Multimodal Human Activity Recognition. arXiv. 20242402.19258 [Google Scholar]
- 59.Sheng B., Han R., Xiao F., Guo Z., Gui L. MetaFormer: Domain-Adaptive WiFi Sensing with Only One Labelled Target Sample. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;8:1–27. doi: 10.1145/3643550. [DOI] [Google Scholar]
- 60.Pan Y., Zhou Z., Gong W., Fang Y. SAT: A Selective Adversarial Training Approach for WiFi-based Human Activity Recognition. IEEE Trans. Mob. Comput. 2024;23:12706–12716. doi: 10.1109/TMC.2024.3420405. [DOI] [Google Scholar]
- 61.Yang J., Zou H., Xie L. SecureSense: Defending Adversarial Attack for Secure Device-Free Human Activity Recognition. IEEE Trans. Mob. Comput. 2024;23:823–834. doi: 10.1109/TMC.2022.3226742. [DOI] [Google Scholar]
- 62.Luo F., Khan S., Jiang B., Wu K. Vision Transformers for Human Activity Recognition using WiFi Channel State Information. IEEE Internet Things J. 2024;11:28111–28122. doi: 10.1109/JIOT.2024.3375337. [DOI] [Google Scholar]
- 63.Zhang Y., Wang G., Liu H., Gong W., Gao F. WiFi-Based Indoor Human Activity Sensing: A Selective Sensing Strategy and a Multi-Level Feature Fusion Approach. IEEE Internet Things J. 2024;11:29335–29347. doi: 10.1109/JIOT.2024.3397708. [DOI] [Google Scholar]
- 64.Niu K., Wang X., Zhang F., Zheng R., Yao Z., Zhang D. Rethinking Doppler Effect for Accurate Velocity Estimation With Commodity WiFi Devices. IEEE J. Sel. Areas Commun. 2022;40:2164–2178. doi: 10.1109/JSAC.2022.3155523. [DOI] [Google Scholar]
- 65.Wu D., Zeng Y., Gao R., Li S., Li Y., Shah R.C., Lu H., Zhang D. WiTraj: Robust Indoor Motion Tracking With WiFi Signals. IEEE Trans. Mob. Comput. 2023;22:3062–3078. doi: 10.1109/TMC.2021.3133114. [DOI] [Google Scholar]
- 66.Li W., Gao R., Xiong J., Zhou J., Wang L., Mao X., Yi E., Zhang D. WiFi-CSI Difference Paradigm: Achieving Efficient Doppler Speed Estimation for Passive Tracking. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;8:1–29. doi: 10.1145/3659608. [DOI] [Google Scholar]
- 67.Zhang G., Zhang D., He Y., Chen J., Zhou F., Chen Y. Multi-Person Passive WiFi Indoor Localization With Intelligent Reflecting Surface. IEEE Trans. Wirel. Commun. 2023;22:6534–6546. doi: 10.1109/TWC.2023.3244369. [DOI] [Google Scholar]
- 68.Zhang G., Zhang D., Deng H., Wu Y., Zhan F., Chen Y. Practical Passive Indoor Localization With Intelligent Reflecting Surface. IEEE Trans. Mob. Comput. 2024;23:12477–12490. doi: 10.1109/TMC.2024.3414861. [DOI] [Google Scholar]
- 69.Fan Y., Zhang F., Wu C., Wang B., Liu K.J.R. RF-Based Indoor Moving Direction Estimation Using a Single Access Point. IEEE Internet Things J. 2022;9:462–473. doi: 10.1109/JIOT.2021.3083669. [DOI] [Google Scholar]
- 70.Chi G., Yang Z., Xu J., Wu C., Zhang J., Liang J., Liu Y. Wi-drone: Wi-fi-based 6-DoF tracking for indoor drone flight control; Proceedings of the 20th Annual International Conference on Mobile Systems, Applications and Services (MobiSys ‘22); Portland, OR, USA. 27 June–1 July 2022; New York, NY, USA: Association for Computing Machinery; 2022. pp. 56–68. [Google Scholar]
- 71.Zeng Y., Wu D., Xiong J., Liu J., Zhang D. MultiSense: Enabling Multi-person Respiration Sensing with Commodity WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020;4:1–29. doi: 10.1145/3411816. [DOI] [Google Scholar]
- 72.Zhang F., Wu C., Wang B., Wu M., Bugos D., Zhang H., Liu K.J.R. SMARS: Sleep Monitoring via Ambient Radio Signals. IEEE Trans. Mob. Comput. 2021;20:217–231. doi: 10.1109/TMC.2019.2939791. [DOI] [Google Scholar]
- 73.Yu B., Wang Y., Niu K., Zeng Y., Gu T., Wang L., Guan C., Zhang D. WiFi-Sleep: Sleep Stage Monitoring Using Commodity Wi-Fi Devices. IEEE Internet Things J. 2021;8:13900–13913. doi: 10.1109/JIOT.2021.3068798. [DOI] [Google Scholar]
- 74.Liu J., Zeng Y., Gu T., Wang L., Zhang D. WiPhone: Smartphone-based Respiration Monitoring Using Ambient Reflected WiFi Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021;5:1–19. doi: 10.1145/3448092. [DOI] [Google Scholar]
- 75.Hu J., Yang J., Ong J.-B., Wang D., Xie L. ResFi: WiFi-Enabled Device-Free Respiration Detection Based on Deep Learning; Proceedings of the 2022 IEEE 17th International Conference on Control & Automation (ICCA); Naples, Italy. 27–30 June 2022; pp. 510–515. [Google Scholar]
- 76.Xie X., Zhang D., Li Y., Hu Y., Sun Q., Chen Y. Robust WiFi Respiration Sensing in the Presence of Interfering Individual. IEEE Trans. Mob. Comput. 2024;23:8447–8462. doi: 10.1109/TMC.2023.3348879. [DOI] [Google Scholar]
- 77.Jiang W., Xue H., Miao C., Wang S., Lin S., Tian C., Murali S., Hu H., Sun Z., Su L. Towards 3D human pose construction using wifi; Proceedings of the 26th Annual International Conference on Mobile Computing and Networking (MobiCom ‘20); London, UK. 21–25 September 2020; New York, NY, USA: Association for Computing Machinery; 2020. pp. 1–14. [Google Scholar]
- 78.Li C., Liu Z., Yao Y., Cao Z., Zhang M., Liu Y. Wi-fi see it all: Generative adversarial network-augmented versatile wi-fi imaging; Proceedings of the 18th Conference on Embedded Networked Sensor Systems (SenSys ‘20); Yokohama, Japan. 16–19 November 2020; New York, NY, USA: Association for Computing Machinery; 2020. pp. 436–448. [Google Scholar]
- 79.Ren Y., Wang Z., Wang Y., Tan S., Chen Y., Yang J. GoPose: 3D Human Pose Estimation Using WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–25. doi: 10.1145/3534605. [DOI] [Google Scholar]
- 80.Pallaprolu A., Korany B., Mostofi Y. Wiffract: A new foundation for RF imaging via edge tracing; Proceedings of the 28th Annual International Conference on Mobile Computing and Networking (MobiCom ‘22); Sydney, NSW, Australia. 17–21 October 2022; New York, NY, USA: Association for Computing Machinery; 2022. pp. 255–267. [Google Scholar]
- 81.Zhou Y., Huang H., Yuan S., Zou H., Xie L., Yang J. MetaFi++: WiFi-Enabled Transformer-Based Human Pose Estimation for Metaverse Avatar Simulation. IEEE Internet Things J. 2023;10:14128–14136. doi: 10.1109/JIOT.2023.3262940. [DOI] [Google Scholar]
- 82.Wang X., Niu K., Yu A., Xiong J., Yao Z., Wang J., Li W., Zhang D. WiMeasure: Millimeter-level Object Size Measurement with Commodity WiFi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2023;7:1–26. doi: 10.1145/3596250. [DOI] [Google Scholar]
- 83.Yin C., Miao X., Chen J., Jiang H., Yang J., Zhou Y., Wu M., Chen Z. PowerSkel: A Device-Free Framework Using CSI Signal for Human Skeleton Estimation in Power Station. IEEE Internet Things J. 2024;11:20165–20177. doi: 10.1109/JIOT.2024.3369856. [DOI] [Google Scholar]
- 84.Yao Z., Wang X., Niu K., Zheng R., Wang J., Zhang D. WiProfile: Unlocking Diffraction Effects for Sub-Centimeter Target Profiling Using Commodity WiFi Devices; Proceedings of the 30th Annual International Conference on Mobile Computing and Networking (ACM MobiCom ‘24); Washington, DC, USA. 18–22 November 2024; New York, NY, USA: Association for Computing Machinery; 2024. pp. 185–199. [Google Scholar]
- 85.Wu D., Zeng Y., Zhang F., Zhang D. WiFi CSI-based device-free sensing: From Fresnel zone model to CSI-ratio model. CCF Trans. Pervasive Comput. Interact. 2022;4:88–102. doi: 10.1007/s42486-021-00077-z. [DOI] [Google Scholar]
- 86.Niu K., Wang X., Yao Z., Zhang F., Cheng S., Jiang Y., Zhang D. How Target’s Location and Orientation Affect Velocity Extraction Accuracy in WiFi Sensing Systems; Proceedings of the ACM Turing Award Celebration Conference-China 2023 (ACM TURC ′23); Wuhan, China. 28–30 July 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 35–36. [Google Scholar]
- 87.Zhang F., Jin B., Zhang D. Ubiquitous Wireless Sensing-Theory, Technique and Application; Proceedings of the ACM Turing Award Celebration Conference-China 2023 (ACM TURC ‘23); Wuhan, China. 28–30 July 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 33–34. [Google Scholar]
- 88.Zhang J.A., Wu K., Huang X., Guo Y.J., Zhang D., Heath R.W. Integration of Radar Sensing into Communications with Asynchronous Transceivers. IEEE Commun. Mag. 2022;60:106–112. doi: 10.1109/MCOM.003.2200096. [DOI] [Google Scholar]
- 89.Zeng Y., Wu D., Xiong J., Zhang D. Boosting WiFi Sensing Performance via CSI Ratio. IEEE Pervasive Comput. 2021;20:62–70. doi: 10.1109/MPRV.2020.3041024. [DOI] [Google Scholar]
- 90.Zeng Y., Wu D., Xiong J., Yi E., Gao R., Zhang D. FarSense: Pushing the Range Limit of WiFi-based Respiration Sensing with CSI Ratio of Two Antennas. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019;3:1–26. doi: 10.1145/3351279. [DOI] [Google Scholar]
- 91.Zeng Y., Liu J., Xiong J., Liu Z., Wu D., Zhang D. Exploring Multiple Antennas for Long-range WiFi Sensing. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;5:1–30. doi: 10.1145/3494979. [DOI] [Google Scholar]
- 92.Li Y., Wu D., Zhang J., Xu X., Xie Y., Gu T., Zhang D. DiverSense: Maximizing Wi-Fi Sensing Range Leveraging Signal Diversity. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–28. doi: 10.1145/3536393. [DOI] [Google Scholar]
- 93.Wang X., Niu K., Xiong J., Qian B., Yao Z., Lou T., Zhang D. Placement Matters: Understanding the Effects of Device Placement for WiFi Sensing. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–25. doi: 10.1145/3517237. [DOI] [Google Scholar]
- 94.Xie Y., Li Z., Li M. Precise Power Delay Profiling with Commodity WiFi; Proceedings of the 21st Annual International Conference on Mobile Computing and Networking (MobiCom ‘15); Paris, France. 7–11 September 2015; New York, NY, USA: Association for Computing Machinery; 2015. pp. 53–64. [Google Scholar]
- 95.Gringoli F., Schulz M., Link J., Hollick M. Free Your CSI: A Channel State Information Extraction Platform For Modern Wi-Fi Chipsets; Proceedings of the 13th International Workshop on Wireless Network Testbeds, Experimental Evaluation & Characterization (WiNTECH ‘19); Los Cabos, Mexico. 25 October 2019; New York, NY, USA: Association for Computing Machinery; 2019. pp. 21–28. [Google Scholar]
- 96.Hernandez S.M., Bulut E. Lightweight and Standalone IoT Based WiFi Sensing for Active Repositioning and Mobility; Proceedings of the 2020 IEEE 21st International Symposium on “A World of Wireless, Mobile and Multimedia Networks” (WoWMoM); Cork, Ireland. 31 August–3 September 2020; pp. 277–286. [Google Scholar]
- 97.Hernandez S.M., Bulut E. WiFi Sensing on the Edge: Signal Processing Techniques and Challenges for Real-World Systems. IEEE Commun. Surv. Tutorials. 2023;25:46–76. doi: 10.1109/COMST.2022.3209144. [DOI] [Google Scholar]
- 98.Gringoli F., Cominelli M., Blanco A., Widmer J. AX-CSI: Enabling CSI Extraction on Commercial 802.11ax Wi-Fi Platforms; Proceedings of the 15th ACM Workshop on Wireless Network Testbeds, Experimental evaluation & CHaracterization (WiNTECH ‘21); New Orleans, LA, USA. 31 January–4 February 2022; New York, NY, USA: Association for Computing Machinery; 2021. pp. 46–53. [Google Scholar]
- 99.Jiang Z., Luan T.H., Ren X., Lv D., Hao H., Wang J., Zhao K., Xi W., Xu Y., Li R. Eliminating the Barriers: Demystifying Wi-Fi Baseband Design and Introducing the PicoScenes Wi-Fi Sensing Platform. IEEE Internet Things J. 2022;9:4476–4496. doi: 10.1109/JIOT.2021.3104666. [DOI] [Google Scholar]
- 100.Yousefi S., Narui H., Dayal S., Ermon S., Valaee S. A Survey on Behavior Recognition Using WiFi Channel State Information. IEEE Commun. Mag. 2017;55:98–104. doi: 10.1109/MCOM.2017.1700082. [DOI] [Google Scholar]
- 101.Ma Y., Zhou G., Wang S., Zhao H., Jung W. SignFi: Sign Language Recognition Using WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;2:1–21. doi: 10.1145/3191755. [DOI] [Google Scholar]
- 102.Palipana S., Rojas D., Agrawal P., Pesch D. FallDeFi: Ubiquitous Fall Detection using Commodity Wi-Fi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;1:1–25. doi: 10.1145/3161183. [DOI] [Google Scholar]
- 103.Guo L., Guo S., Wang L., Lin C., Liu J., Lu B., Fang J., Liu Z., Shan Z., Yang J. Wiar: A Public Dataset for Wifi-Based Activity Recognition. IEEE Access. 2019;7:154935–154945. doi: 10.1109/ACCESS.2019.2947024. [DOI] [Google Scholar]
- 104.Zheng Y., Zhang Y., Qian K., Zhang G., Liu Y., Wu C., Yang Z. Zero-Effort Cross-Domain Gesture Recognition with Wi-Fi; Proceedings of the 17th Annual International Conference on Mobile Systems, Applications and Services (MobiSys ′19); Seoul, Republic of Korea. 17–21 June 2019; New York, NY, USA: Association for Computing Machinery; 2019. pp. 313–325. [Google Scholar]
- 105.Xiao R., Liu J., Han J., Ren K. OneFi: One-Shot Recognition for Unseen Gesture via COTS WiFi; Proceedings of the 19th ACM Conference on Embedded Networked Sensor Systems (SenSys ‘21); Coimbra, Portugal. 15–17 November 2021; New York, NY, USA: Association for Computing Machinery; 2021. pp. 206–219. [Google Scholar]
- 106.Yang J., Chen X., Zou H., Lu X., Wang D., Yang S.J., Huang H., Zhou Y., Chen X., Xu Y., et al. MM-Fi: Multi-Modal Non-Intrusive 4D Human Dataset for Versatile Wireless Sensing. In: Oh A., Naumann T., Globerson A., Saenko K., Hardt M., Levine S., editors. Advances in Neural Information Processing Systems. Volume 36. Curran Associates, Inc.; Red Hook, NY, USA: 2023. pp. 18756–18768. [Google Scholar]
- 107.Xie S., Xie L. SenseFi: A library and benchmark on deep-learning-empowered WiFi human sensing. Patterns. 2023;4:100703. doi: 10.1016/j.patter.2023.100703. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108.Cominelli M., Gringoli F., Restuccia F. Exposing the CSI: A Systematic Investigation of CSI-based Wi-Fi Sensing Capabilities and Limitations; Proceedings of the 2023 IEEE International Conference on Pervasive Computing and Communications (PerCom); Atlanta, GA, USA. 13–17 March 2023; pp. 81–90. [Google Scholar]
- 109.Yi E., Zhang F., Xiong J., Niu K., Yao Z., Zhang D. Enabling WiFi Sensing on New-generation WiFi Cards. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;7:1–26. doi: 10.1145/3633807. [DOI] [Google Scholar]
- 110.Yang Z., Zhang Y., Qian K., Wu C. SLNet: A Spectrogram Learning Neural Network for Deep Wireless Sensing; Proceedings of the 20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23); Boston, MA, USA. 17–19 April 2023; Berkeley, CA, USA: USENIX Association; 2023. pp. 1221–1236. [Google Scholar]
- 111.Chi G., Yang Z., Wu C., Xu J., Gao Y., Liu Y., Han T.X. RF-Diffusion: Radio Signal Generation via Time-Frequency Diffusion; Proceedings of the 30th Annual International Conference on Mobile Computing and Networking (ACM MobiCom ‘24); Washington, DC, USA. 18–22 November 2024; New York, NY, USA: Association for Computing Machinery; 2024. pp. 77–92. [Google Scholar]
- 112.Hou W., Wu C. RFBoost: Understanding and Boosting Deep WiFi Sensing via Physical Data Augmentation. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;8:1–26. doi: 10.1145/3659620. [DOI] [Google Scholar]
- 113.He Y., Liu J., Li M., Yu G., Han J. Forward-Compatible Integrated Sensing and Communication for WiFi. IEEE J. Sel. Areas Commun. 2024;42:2440–2456. doi: 10.1109/JSAC.2024.3413955. [DOI] [Google Scholar]
- 114.Wu C., Huang X., Huang J., Xing G. Enabling Ubiquitous WiFi Sensing with Beamforming Reports; Proceedings of the ACM SIGCOMM 2023 Conference (ACM SIGCOMM ‘23); New York, NY, USA. 10 September 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 20–32. [Google Scholar]
- 115.Yi E., Wu D., Xiong J., Zhang F., Niu K., Li W., Zhang D. BFMSense: WiFi Sensing Using Beamforming Feedback Matrix; Proceedings of the 21st USENIX Symposium on Networked Systems Design and Implementation (NSDI24); Santa Clara, CA, USA. 16–18 April 2024; Berkeley, CA, USA: USENIX Association; 2024. pp. 1697–1712. [Google Scholar]
- 116.Korany B., Karanam C.R., Cai H., Mostofi Y. XModal-ID: Using WiFi for Through-Wall Person Identification from Candidate Video Footage; Proceedings of the 25th Annual International Conference on Mobile Computing and Networking (MobiCom ‘19); Los Cabos, Mexico. 21–25 October 2019; New York, NY, USA: Association for Computing Machinery; 2019. pp. 1–15. [Google Scholar]
- 117.De Sanctis M., Domenico S.D., Fioravanti D., Abellan E.B., Rossi T., Cianca E. RF-Based Device-Free Counting of People Waiting in Line: A Modular Approach. IEEE Trans. Veh. Technol. 2022;71:10471–10484. doi: 10.1109/TVT.2022.3182548. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Not applicable.




