Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2024 Nov 10;24(22):7195. doi: 10.3390/s24227195

Commodity Wi-Fi-Based Wireless Sensing Advancements over the Past Five Years

Hai Zhu 1,*, Enlai Dong 1, Mengmeng Xu 1, Hongxiang Lv 1, Fei Wu 1
Editor: Stephan Sand1
PMCID: PMC11597943  PMID: 39598974

Abstract

With the compelling popularity of integrated sensing and communication (ISAC), Wi-Fi sensing has drawn increasing attention in recent years. Starting from 2010, Wi-Fi channel state information (CSI)-based wireless sensing has enabled various exciting applications such as indoor localization, target imaging, activity recognition, and vital sign monitoring. In this paper, we retrospect the latest achievements of Wi-Fi sensing using commodity-off-the-shelf (COTS) devices from the past 5 years in detail. Specifically, this paper first presents the background of the CSI signal and related sensing models. Then, recent studies are categorized from two perspectives, i.e., according to their application scenario diversity and the corresponding sensing methodology difference, respectively. Next, this paper points out the challenges faced by Wi-Fi sensing, including domain dependency and sensing range limitation. Finally, three imperative research directions are highlighted, which are critical for realizing more ubiquitous and practical Wi-Fi sensing in real-life applications.

Keywords: Wi-Fi sensing, CSI, commodity-off-the-shelf, integrated sensing and communication

1. Introduction

The demand for ubiquitous internet connection has catalyzed the vast deployment of Wi-Fi infrastructures over the past decades, making Wi-Fi signal available almost everywhere. With the rapid progress of wireless communication and signal processing techniques, researchers have successfully reused Wi-Fi as a sensing platform beyond its traditional use as a pure communication medium, which further gives birth to the idea of integrated sensing and communication (ISAC) with Wi-Fi [1,2,3]. After years of persistent research, Wi-Fi sensing is drawing huge attention from both academia and industry [4]. Both communities recognize ISAC as a compelling technology capable of improving spectrum efficiency and reducing the hardware cost [5]. It is worth mentioning that, starting from 2020, the IEEE 802.11 working group established an IEEE 802.11bf standardization group for encompassing wireless sensing within the new version of 802.11 standard, turning Wi-Fi sensing into reality.

The basic rational behind Wi-Fi sensing is quite straightforward [6]. When wireless signal propagates from the transmitter to the receiver through multiple paths, a phenomenon called multi-path effect occurs, whereby the superimposed receiving signal intrinsically contains the signal component reflected or diffracted by the sensing target. Therefore, by analyzing the target “modulated” receiving signals, researchers can recover the rich information regarding the target, such as location and activity. Compared with classic sensor-based and vision-based sensing paradigms, Wi-Fi wireless sensing has the advantages of low-cost ubiquity, wide coverage, non-intrusiveness, and privacy-protection. Due to its appealing superiority, numerous Wi-Fi sensing applications have been developed, ranging from coarse-grained motion detection [7] and activity recognition [8] to fine-grained localization [9] and breath monitoring [10].

Inspired by existing survey papers [11,12,13,14,15], this paper investigates the thrilling achievements made within the last 5 years and presents an in-depth analysis of these sensing systems, aiming to facilitate further research in the Wi-Fi sensing field. This paper first divides existing works according to different application scenarios, including localization and tracking, activity recognition, vital sign monitoring, and target imaging. For each category, both application-specific problems and solutions are compared and summarized. Then, this paper further classifies recent studies based on the methodology adopted, whether it is model-based, handcrafted pattern extraction-based, or deep learning-based, pointing out the pros and cons of each method. Furthermore, this paper highlights the remaining challenges of current works such as generalization issues and large-scale perception. Future research directions and the need for further study are discussed in the end. The main contributions of this work are summarized as follows:

  • To the best of our knowledge, this is the latest comprehensive survey in the Wi-Fi sensing field, covering the greatest and most recent progresses made over the past 5 years.

  • We categorize existing studies from two distinct perspectives, i.e., application-based and methodology-based, and present an in-depth analysis of recent works.

  • We highlight the key challenges encountered in existing studies and present a thorough discussion about three promising research directions for Wi-Fi sensing.

The rest of this paper is organized as follows. In Section 2, we briefly introduce the concept of channel state information (CSI) and explain several popular sensing models. In Section 3, we classify state-of-the-art works with regard to two criteria, i.e., application variety and methodology difference. Practical limitations and challenges are analyzed in Section 4. In Section 5, a detailed discussion about future trends in Wi-Fi sensing is provided. Finally, we conclude this article in Section 6.

2. Preliminary

Before analyzing Wi-Fi sensing, we briefly introduce the necessary background of channel state information (CSI) and several general signal sensing models.

2.1. Channel State Information

Serving as a key metric of a communication system, CSI depicts how a signal propagates through a wireless channel. Indeed, a wireless communication channel can be defined as follows:

Y = H × X + N (1)

where X and Y are the transmitted and received signal, respectively. H is the channel matrix representing CSI and N denotes the channel noise.

In a typical indoor environment, shown in Figure 1, a signal sent by the transmitter (Tx) travels through multiple paths before arriving at the receiver (Rx), which is also known as the multi-path effect. Therefore, assuming there are L different paths, the wireless channel H can be mathematically expressed as channel impulse response (CIR) [6], as follows:

ht=i=1Laiejθiδtτi (2)

where ai, θi, and τi are the complex amplitude attenuation, phase shift, and propagation time delay of the i-th path, respectively; δt is the Dirac delta function. Each impulse in the summation of Equation (2) represents a delayed multi-path component, multiplied by its corresponding amplitude and phase variation.

Figure 1.

Figure 1

Typical indoor multi-path Wi-Fi propagation.

As shown in Figure 1, when a person moves inside the signal zone, the human body will inevitably alter the specific propagation path, thus changing the CIR. Hence, the underlying principle of wireless sensing is analyzing human-induced channel variation. However, CIR cannot be precisely measured with commodity Wi-Fi devices, especially given the limited bandwidth of Wi-Fi. Fortunately, with the adoption of the orthogonal frequency division multiplex (OFDM) technique in present IEEE 802.11 standard, researchers resorted to studying channel frequency response (CFR), an equivalent channel representation of CIR in the frequency domain.

CFRf=CFRfejCFRf (3)

where CFRf and CFRf represent the amplitude–frequency and phase–frequency response of CFR, respectively. With proper driver modifications, researchers have been able to obtain an OFDM-based sampling version of CFR with a commercial-off-the-shelf (COTS) Wi-Fi network interface card (NIC) since 2010 [16,17], greatly prompting the development of Wi-Fi sensing [12]. To be specific, the extracted CFR depicts the amplitude and phase of different subcarriers, which can be expressed as follows:

Hfi=HfiejHfi (4)

where Hfi is the CFR sampled at the i-th subcarrier with the central frequency of fi. In fact, the CSI data H=Hfi|i1, N used in most research papers refers directly to the definition given by Equation (4), i.e., a sampled version of CFR at the granularity of a subcarrier level.

Generally speaking, this sampled CFR lays the foundations for advanced Wi-Fi sensing, paving the way for the feasibility of various modern applications. CSI data contains rich information on signal propagation, and we will use CSI to simply signify the raw Wi-Fi data for brevity in the following part.

2.2. Signal Sensing Models

2.2.1. Fresnel Zone-Based Reflection Model

Taking one pair of the Tx-Rxlink as an example, Fresnel zones are concentric ellipses with two foci corresponding to the Tx and Rx, as P1 and P2, shown in Figure 2. For a given radio length λ, the n-th Fresnel zone boundary containing n ellipses can be defined as follows:

PiQn+QnP2P1P2=nλ/2 (5)

where Qn is a point on the n-th Fresnel zone boundary. The n-th Fresnel zone refers to the elliptic annulus between the (n−1)-th and n-th ellipse boundary, while the innermost ellipse is called the first Fresnel zone (FFZ). Equation (5) indicates that the path length of the signal reflected through the n-th Fresnel zone boundary is nλ/2 is longer than that of the Line-of-Sight (LOS) path, i.e., P1P2.

Figure 2.

Figure 2

Geometry of Fresnel zone reflection sensing [18].

The Fresnel zone-based reflection model [18] characterizes how the amplitude and phase of CSI change when a target moves outside the FFZ. The key property of the reflection sensing model is when a target moves across a series of Fresnel zone boundaries, CSI amplitude and phase will show a continuous sinusoidal-like pattern, which can be utilized for sensing applications such as respiration and walking direction detection [19].

2.2.2. Fresnel Zone-Based Diffraction Model

According to the RF propagation theory, more than 70% of the signal energy is transferred via the FFZ. Therefore, when a target moves inside the FFZ, signal diffraction becomes more important and dominates the received signal variation. As shown in Figure 3, the Fresnel zone-based diffraction model [20] depicts how the amplitude and phase of CSI change when a target moves inside the FFZ. The key property is when sensing activity inside the FFZ, the CSI amplitude variation will show different shapes, be it either monotonically decreased or non-monotonous, “W”, depending on the target size. Apart from respiration monitoring, the diffraction sensing model has also been proved effective for recognizing exercise and daily exercise [8].

Figure 3.

Figure 3

Geometry of Fresnel zone diffraction sensing [20].

2.2.3. Scattering Sensing Model

One main limitation of the previous models is that the simple reflection or diffraction assumption may not hold true when considering complex target motions, in cases where signals are scattered from multiple human body parts. Different from the Fresnel zone-based model, the scattering sensing model treats all objects as scatters, taking into account all multi-paths together. As marked as red circles in Figure 4, intuitively, the scattering model considers each scatter point as a virtual Tx, e.g., the static walls, and the arm and leg of the moving human. Given numerous multi-paths are considered, the scattering model is in fact a statistical model generally applicable to complex indoor scenarios. The scattering sensing model has been adopted in various speed-oriented tasks [21,22], achieving robust performance even with non-line-of-sight (NLOS) occlusion.

Figure 4.

Figure 4

Signal scattering sensing model.

3. Wi-Fi Sensing

Serving as a key property in future wireless systems, Wi-Fi sensing has enabled various important applications. In this section, we categorize recent works based on two aspects, i.e., whether they are application-oriented or methodology-oriented. Since there are quite a few references in this section, for the reader’s convenience, we provide an index of all mentioned references in Table A1 of Appendix A.

3.1. Wi-Fi Sensing Applications

In this section, we divide the related works into seven categories, i.e., presence detection, gait recognition, gesture recognition, activity recognition, localization and tracking, vital sign monitoring, and pose construction and imaging, as shown in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7. In each table, “Application” implies detailed application demand, “User number” signifies the number of sensing targets supported by the study, “Device type” indicates the specific sensing equipment used, and “NLOS” shows whether the sensing system can work in a non-line-of-sight scenario or not.

Table 1.

Presence detection.

Year Reference Application Performance User Number Device Type NLOS
2022 WiCPD [23] In-car child presence detection 96.56–100% real-time detection rate 1 NXP Wi-Fi chipset Y
2023 Hu et al. [24] Proximity detection 95% and 99% true positive rate for distance-based and room-based detection 1 NXP Wi-Fi chipset Y
2024 Zhu et al. [25] Human and non-human differentiation 95.57% average accuracy 1 human or pet COTS device Y
2024 WI-MOID [26] Edge device-based human and non-human differentiation 97.34% accuracy and 1.75% false alarm rate 1 human or non-human subject Wi-Fi edge device Y

Presence detection. Presence detection determines whether a target exists or not within the sensing area and serves as the prerequisite for further sensing tasks. Target presence detection could enable many modern applications, such as security systems and smart homes. Although usually included as a detector module in most studies, there have been some new applications based on presence detection. As shown in Table 1, WiCPD [23] studied child presence detection in a smart car scenario, preventing potential harm to children if left alone in a vehicle. Hu et al. [24] considered target location relative to the sensing device, supporting more intelligent control systems using this area-aware context. In addition, Zhu et al. [25] and WI-MOID [26] further differentiated human from non-human targets to mitigate influence from unwanted objects, avoiding unnecessary false alarms.

Table 2.

Gait recognition.

Year Reference Application Performance User Number Device Type NLOS
2021 GaitSense [27] Gait-based human identification 93.2% for 5 users and 76.2% for 11 users 11 Intel 5300 N
2021 GaitWay [28] Gait speed estimation 0.12 m median error 1 Intel 5300 Y
2022 CAUTION [29] Gait-based human authentication 93.06 average accuracy 15 TP-Link N750 router N
2022 Wi-PIGR [30] Gait recognition 93.5% for single user and 77.15% for 50 users 1–50 Intel 5300 N
2023 Auto-Fi [31] Gesture and gait recognition 86.83% for gesture; 79.61% for gait 1 Atheros chipset N
2023 GaitFi [32] Gait recognition 94.2% accuracy 12 TP-Link N750 router N
2024 Wi-Diag [33] Multi-subject abnormal gait diagnosis 87.77% average accuracy 4 Intel 5300 N

Gait recognition. Gait, a unique biomarker, refers to the distinctive walking character of different people and has been used for human identification and authentication applications. Early gait sensing works usually required users to walk on fixed trajectories within restricted areas, while recent studies, e.g., GaitSense [27], GaitWay [28], and Wi-PIGR [30], aimed for path independent gait recognition where users can walk along arbitrary paths even in a through-the-wall scenario. In addition, CAUTION [29], Auto-Fi [31], and GaitFi [32] tried to realize robust gait recognition with limited training data, while Wi-Diag [33] further studied more challenging multi-human recognition problems. As depicted in Table 2, all these works greatly contribute to more ubiquitous gait-based sensing applications.

Table 3.

Gesture recognition.

Year Reference Application Performance User Number Device Type NLOS
2021 Kang et al. [34] Gesture recognition 3–12.7% improvement 1 Widar Dataset N
2021 WiGesture [35] Gesture recognition 92.8–94.5% accuracy 1 Intel 5300 N
2022 HandGest [36] Handwriting recognition 95% accuracy 1 Intel 5300 N
2022 DPSense-WiGesture [37] Gesture recognition 94% average accuracy 1 Intel 5300 N
2022 Niu et al. [38] Gesture recognition 96% accuracy 1 Intel 5300 Y
2022 Widar 3.0 [39] Cross-domain gesture recognition 92.7% in-domain and 82.6–92.4% cross-domain accuracy 1 Intel 5300 N
2022 WiFine [40] Gesture recognition 96.03% accuracy in 0.19 s 1 Raspberry Pi 4B N
2023 UniFi [41] Gesture recognition 99% and 90–98% accuracy for in-domain and cross-domain recognition 1 Widar dataset N
2023 WiTransformer [42] Gesture recognition 86.16% accuracy 1 Widar dataset N
2024 AirFi [43] Gesture recognition 90% accuracy 1 TP-Link N750 router N
2024 WiCGesture [44] Continuous gesture recognition 89.6% for digits and 88.3% for Greek letters 1 Intel 5300 N

Gesture recognition. Wireless gesture recognition has emerged as an important part of modern human computer interaction, enabling wide applications including smart home control and virtual reality. Previous studies tried to learn the intricate pattern between signal variation and human gesture under the one-to-one mapping assumption. However, this assumption does not hold, since the received signal is highly dependent on the relative location and orientation of users, as proven by the Fresnel reflection model [18]. Thus, recent works mainly focused on realizing a position-independent robust gesture recognition system, as illustrated in Table 3. Kang et al. [34], Widar 3.0 [39], UniFi [41], WiTransformer [42], and AirFi [43] leverage various deep learning methods, e.g., adversarial learning, multi-view network, and few-shot learning, to realize a robust and efficient recognition. On the other hand, WiGesture [35], HandGest [36], DPSense-WiGesture [37], Niu et al. [38], and WiCGesture [44] attempted to extract distinct and consistent features from a hand-oriented perspective, realizing reliable and continuous recognition either through more fine-grained signal segmentation or signal quality assessment. In addition, WiFine [40] managed to realize real-time gesture recognition using low-end edge devices, e.g., Raspberry Pi. Overall, these methods bring Wi-Fi gesture recognition one step closer to more practical uses.

Table 4.

Activity recognition.

Year Reference Application Performance User Number Device Type NLOS
2020 Wang et al. [45] People counting and recognition 86% average accuracy 4 COTS devices N
2021 Ma et al. [46] Activity recognition 97% average accuracy 1 Intel 5300 N
2021 MCBAR [47] Activity recognition 90% average accuracy 1 Atheros chipset N
2021 WiMonitor [48] Location and activity monitoring Not applicable 1 Intel 5300 Y
2022 DeFall [49] Fall detection 95% detection rate and 1.5% false alarm rate 1 Intel 5300 Y
2022 Ding et al. [50] Activity recognition 96.85% average accuracy 1 Intel 5300 N
2022 EfficientFi [51] Activity recognition 98% accuracy 1 TP-Link N750 router N
2022 TOSS [52] Activity recognition 82.69% average accuracy 1 Intel 5300 N
2023 FallDar [53] Fall detection 5.7% false alarm rate and 3.4% missed alarm rate 1 Intel 5300 Y
2023 SHARP [54] Activity recognition 95% average accuracy 1 ASUS RT-AC86U router N
2023 Liu et al. [55] Moving receiver-based activity recognition 10°, 1 cm and 98% accuracy for direction, displacement, and activity estimation 1 COTS WiFi 6 device N
2023 WiCross [56] Target passing detection 95% accuracy 1 Intel 5300 N
2024 i-Sample [57] Activity recognition 10% accuracy gain 1 Intel 5300 N
2024 MaskFi [58] Activity recognition 97.61% average accuracy 1 TP-Link N750 router N
2024 MetaFormer [59] Activity recognition Improved accuracy in various cross-domain scenarios 1 SiFi, Widar, Wiar datasets N
2024 SAT [60] Activity recognition Improved accuracy and robustness 1 Intel 5300 N
2024 SecureSense [61] Activity recognition under adversarial attack Robust performance under various attacks 1 TP-Link N750 router N
2024 Luo et al. [62] Activity recognition 98.78% accuracy 1 UT-HAR dataset N
2024 WiSMLF [63] Activity recognition 92% average accuracy 1 Intel 5300 N

Activity recognition. Wi-Fi-based human activity recognition (HAR) has become the most studied research topic over the past years, covering many applications including people counting [45], fall detection [49,53], door-passing detection [56], and daily activities. Table 4 shows the summary of recent HAR works. Most works tried to address performance degradation due to location, person, and environment dynamics, also known as domain-dependent problems [46,47,50,52,54,57,58,59,62,63]. In addition, WiMonitor [48] studied continuous long-term human activity monitoring, capturing user information such as location change, activity intensity, and time. Moreover, EfficientFi [51] considered the signal transfer-induced communication problem in a large-scale sensing scenario, providing a cloud-enabled solution with efficient CSI compression, while SAT [60] and SecureSense [61] proposed robust sensing schemes under various adversarial attacks. Liu et al. [55] proposed a dynamic Fresnel zone sensing model using a moving receiver such as a smartphone, filling the gap in existing fixed-location transceivers.

Table 5.

Localization and tracking.

Year Reference Application Performance User Number Device Type NLOS
2022 Niu et al. [64] Velocity estimation-based tracing 9.38 cm/s, 13.42° and 31.08 cm median error in speed, heading and location estimation 1 Intel 5300 Y
2023 WiTraj [65] Human walking tracking 2.5% median tracking error 1 Intel 5300 N
2024 FewSense [66] Tracking 34 cm median error 1 Intel 5300 N
2023 Zhang et al. [67] Multi-person localization Sub-centimeter accuracy 1–3 COTS WiFi device + IRS N
2024 Zhang et al. [68] Passive localization 0.11 m average error 1 VNA N
2022 Fan et al. [69] Moving direction estimation 6.9° median error for moving direction estimation; 16.6° mean error for rotation angle estimation 1 Atheros chipset Y
2022 Wi-Drone [70] Tracking-based indoor drone flight control 26.1 cm average location accuracy and 3.8° rotation accuracy 1 AR9580 NICs N

Localization and tracking. Due to the limited channel bandwidth and antenna number of COTS Wi-Fi devices, there have not been many studies on Wi-Fi-based localization and tracking, as shown in Table 5. Recent works tried to improve tracking performance through more accurate target velocity estimations using a moving-induced Doppler Frequency Shift (DFS). Niu et al. [64] optimized velocity estimation by devising a dynamic selection scheme, which can choose the optimal set of receivers for tracking. To better track human walking, WiTraj [65] intelligently combined multi-view information provided by different receivers and differentiated walking with in-place activity to avoid tracking error accumulation. FewSense [66] creatively fused phase and information for better DFS estimation, achieving high accuracy even with fewer CSI samples. In addition to these works, Zhang et al. [67,68] achieved sub-centimeter localization accuracy using the intelligent reflecting surface (IRS) technique. By constructing an IRS, researchers can modulate the spatial distribution of the Wi-Fi signal, improving the spatial resolution of Wi-Fi localization. While promising, their current prototype systems are realized using a vector network analyzer (VNA), requiring further study with a COTS device. Apart from the device-free tracking mentioned above, Fan et al. [69] and Wi-Drone [70] studied device-based tracking applications. Fan et al. [69] obtained accurate moving direction and in-place rotation angle estimation using a single access point, while Wi-Drone [70] realized the first Wi-Fi tracking-based indoor drone flight control system, providing promising possible solutions for indoor localization and navigation.

Table 6.

Vital sign monitoring.

Year Reference Application Performance User Number Device Type NLOS
2020 MultiSense [71] Multi-person respiration sensing 0.73 bpm mean error 4 Intel 5300 Y
2021 SMARS [72] Breath estimation and sleep stage recognition 0.47 bpm median error and 88% accuracy 1 Atheros chipset Y
2021 WiFi-Sleep [73] Sleep stage monitoring 81.8% accuracy 1 Intel 5300 N
2021 WiPhone [74] Respiration monitoring 0.31 bpm average error 1 ASUS RT-AC86U router and Google Nexus 5 smartphone Y
2022 ResFi [75] Respiration detection 96.05% accuracy 1 ASUS RT-AC86U router N
2024 Xie et al. [76] Respiration sensing with interfering individual 32% mean absolute error reduction 1 VNA or Intel 5300 N

Vital sign monitoring. Vital signs play a crucial role in monitoring people’s health and well-being, providing useful information for early prediction and interference with potential diseases. As shown in Table 6, CSI-based vital sign detection mainly focused on respiration estimation. MultiSense [71] studied the multi-person respiration sensing problem, while SMARS [72] and WiFi-Sleep [73] integrated breath monitoring into users’ sleep quality assessment. WiPhone [74] presented a smartphone-based sensing system, achieving robust performance in NLOS scenarios. Xie et al. [76] addressed the motion interference from nearby individuals, bringing respiration monitoring closer to practical application.

Table 7.

Pose construction and imaging.

Year Reference Application Performance User Number Device Type NLOS
2020 WiPose [77] Pose construction 2.83 cm average error 1 Intel 5300 N
2020 WiSIA [78] Target imaging Not applicable 1 Intel 5300 N
2022 GoPose [79] 3D human pose estimation 4.7 cm accuracy 1 or 2 Intel 5300 Y
2022 Wiffract [80] Still object imaging 86.7% letter reading accuracy 1 Intel 5300 Y
2023 MetaFi++ [81] Pose estimation 97.3% for PCK@50 1 TP-Link N750 router N
2023 WiMeasure [82] Object size measurement 2.6 mm median error 1 Intel 5300 N
2024 PowerSkel [83] Pose estimation 96.27% for PCK@50 1 ESP 32 IoT SoC N
2024 WiProfile [84] 2D target Profiling 1 cm median absolute error 1 target with proper size range Intel 5300 N

Pose construction and imaging. Wi-Fi-based pose estimation and target imaging provides a complementary solution to traditional camera-based perception. As listed in Table 7, WiPose [77], GoPose [79], MetaFi++ [81], and PowerSkel [83] proposed different 3D human skeleton construction frameworks, while WiSIA [78], Wiffract [80], and WiProfile [84] further investigated how to recover target images with Wi-Fi signals. Alternatively, WiMeasure [82] realized millimeter-level high-precision target size measurements, making up for a missing piece of Wi-Fi sensing. It should be noted that in order to achieve fine-grained imaging, the deployment of a high sampling rate and even a customized antenna are usually required, as shown in the subsequent tables. Therefore, Wi-Fi imaging is only applicable for specific application scenarios for the time being.

3.2. Wi-Fi Sensing Methodologies

In this section, we divide the related works into three categories, i.e., model-based sensing, hand-crafted statistical pattern extraction-based sensing, and automatic deep pattern extraction-based sensing, as shown from Table 8, Table 9 and Table 10. In each table, “Methodology” briefly describes the specific method adopted, and “Base signal” refers to the sensing signal constructed with raw CSI, including autocorrelation function (ACF), power spectrum density (PSD), Doppler frequency shift (DFS), body-coordinate velocity profile (BVP), and so on. In addition, “Setting” specifies the signal sampling rate required, the number of Tx-Rx pair used, and certain device settings used for the system implementation and performance evaluation.

Model-based sensing. Since model-based sensing methods have the clear advantage of interpretability, researchers have developed several models for describing the physical relationship between CSI variation and target behavior, detailed in Section 2. As shown in Table 8, the scattering model has been widely used for velocity and periodic pattern extraction [28,49,72], while the diffraction model has been adopted in near-the-LOS scenarios, i.e., within FFZ, for fine-grained sensing tasks [56,80,82,84]. Although less prevalent in Table 8 [55], the Fresnel zone-based reflection model is in fact the most used model. The reflection model is commonly implicitly incorporated in various sensing systems for quantitatively analyzing signal variations and identifying sensing limitations, thus guiding the implementation of more stable and reliable sensing systems [85,86,87].

Table 8.

Model-based sensing.

Year Reference Methodology Performance Base Signal Sensing Range Setting
2021 GaitWay [28] Scattering model 0.12 m median error ACF of CSI 20 m × 23 m 1500 Hz; single pair of Tx-Rx
2021 SMARS [72] Scattering model 0.47 bpm median error and 88% accuracy ACF of CSI 10 m 30 Hz; single pair of Tx-Rx
2022 DeFall [49] Scattering model 95% detection rate and 1.5% false alarm rate ACF of CSI Multi-room 1500 Hz; single pair of Tx-Rx
2022 Wiffract [80] Keller’s Geometrical Theory of Diffraction 86.7% letter reading accuracy Power of CSI 1.5 m Two pairs of Tx-Rx; two-dimensional RX grid synthesis
2023 Liu et al. [55] Dynamic Fresnel zone model 10°, 1 cm and 98% accuracy for direction, displacement and activity estimation CSI Single room 100 Hz; single pair of Tx-Rx
2023 WiCross [56] Diffraction model-based phase pattern extraction 95% accuracy CSI ratio 1 m 1000 Hz; single pair of Tx-Rx
2023 WiMeasure [82] Diffraction model 2.6 mm median error CSI ratio Near the LOS path 500 Hz; three pairs of Tx-Rx
2024 WiProfile [84] Diffraction effect-based profiling + inverse Fresnel transform 1 cm median absolute error CSI 1.5 m × 1 m 500 Hz; single pair of Tx-Rx; One reference receiving antenna connected to Rx via feeder line

Hand-crafted statistical pattern extraction-based sensing. Derived from feature engineering in traditional machine learning processes, researchers have come up with various task-oriented feature extraction schemes, utilizing the in-depth analysis of activity characteristics and advanced signal processing techniques. As shown in Table 9, along with signal processing such as signal segmentation and signal energy estimation, statistical features, such as Doppler frequency shift and speed estimation, motion navigation primitive (MNP), dynamic phase vector (DPV) and motion rotation variable (MRV), have been derived for various sensing tasks. Albeit promising, since feature extraction and selection plays a key role in system performance, hand-crafted features are usually task-specific and not reusable for new tasks, hindering their usage for ubiquitous sensing.

Table 9.

Hand-crafted statistical pattern-based sensing.

Year Reference Methodology Performance Base Signal Sensing Range Setting
2020 MultiSense [71] ICA-based BSS 0.73 bpm mean error Constructed reference-CSI-based signal ratio 4 m × 7.5 m 200 Hz; single pair of Tx-Rx
2020 Wang et al. [45] Statistical pattern analysis 86% accuracy PSD of CSI 3.5 m 10 Hz; single pair of Tx-Rx
2021 WiGesture [35] MNP feature extraction 92.8–94.5% accuracy CSI ratio 4 m × 7 m 400 Hz; two pairs of Tx-Rx
2021 WiMonitor [48] Doppler frequency and activity intensity pattern extraction Not applicable CSI ratio Multi-room 200 Hz; single pair of Tx-Rx
2021 WiPhone [74] Ambient reflection-based pattern extraction 0.31 bpm average error CSI amplitude Multi-room apartment 50 Hz; single pair of Tx-Rx with LOS blocked
2022 HandGest [36] Hand-centric feature extraction, i.e., DPV and MRV 4.7 cm accuracy CSI ratio 1 m 500 Hz; two pairs of Tx-Rx
2022 Niu et al. [64] DFS-based velocity estimation + receiver selection 96.05% accuracy CSI ratio 7 m × 9.8 m 1000 Hz; six pairs of Tx-Rx
2022 Fan et al. [69] 2D-antenna array-based signal parameter estimation 6.9° median error for moving direction estimation; 16.6° mean error for rotation angle estimation Time-reversal resonating strength of CSI 28 m × 36.5 m 200 Hz; single pair of Tx-Rx; half octagonal array of 6 antennas
2022 Wi-Drone [70] Rigid-body coordinate transformation-based absolute pose and relative motion estimation 26.1 cm average location accuracy and 3.8° rotation accuracy CSI 32 m × 18 m Four pairs of Tx-Rx
2022 DPSense-WiGesture [37] Signal segmentation + sensing quality-based signal processing 94% average accuracy CSI 1.2 m 400 Hz; two pairs of Tx-Rx
2022 Niu et al. [38] Position-independent feature extraction, i.e., movement fragment and relative motion direction change 96% accuracy CSI ratio 2 m × 2 m 1000 Hz; 2 pairs of Tx-Rx
2022 WiCPD [23] Feature-based motion, stationary and transition target detector 96.56–100% real-time detection rate ACF of CSI Car 30 Hz; single pair of Tx-Rx
2023 Hu et al. [24] Sub-carrier correlation and covariance feature extraction 95% and 99% true positive rate for distance-based and room-based detection Power of CSI Multi-room 30 Hz; single pair of Tx-Rx
2023 WiTraj [65] DFS extraction + multi-view trajectory estimation + motion detection 2.5% median tracking error CSI ratio 7 m × 6 m 400 Hz; three pairs of Tx-Rx
2023 Zhang et al. [67] Intelligent reflecting surface construction Sub-centimeter accuracy Received signal power 6 m × 6 m Single pair of Tx-Rx
2024 Zhang et al. [68] Intelligent reflecting surface construction 0.11 m average error Received signal power 3 m × 3 m Single pair of Tx-Rx
2024 Xie et al. [76] Respiratory energy-based interference detection and convex optimization-based beam control 32% mean absolute error reduction CSI 9 m × 6 m Single pair of Tx-Rx
2024 WiCGesture [44] Meta motion-based signal segmentation and back-tracking searching-based identification 89.6% for digits and 88.3% for Greek letters CSI ratio 1 m 400 Hz; Two pairs of Tx-Rx
2024 FewSense [66] TD-CSI-based Doppler speed estimation 34 cm median error Time domain CSI difference 7 m × 7 m 1000 Hz; Two pairs of Tx-Rx
2024 WI-MOID [26] Physical and statistical pattern extraction + SVM + state machine 97.34% accuracy and 1.75% false alarm rate ACF of CSI Multi-room 1500 Hz; single pair of Tx-Rx

Automatic deep pattern extraction-based sensing. Since it is challenging to devise effective sensing features, more and more studies have begun leveraging various deep learning models for better accuracy and robustness, such as the Convolution Neural Network (CNN) and Recurrent Neural Network (RNN). As seen in Table 10, the combination of CNN and RNN has been widely adopted in recent works [27,30,32,39,73,77,79] due to its advantage in extracting spatial-temporal features from the CSI signal automatically. In addition, to gain more general representation learning, adversarial learning and few-shot learning have also been used for efficient and robust feature training [29,31,34,43,53,57,60,61]. The end-to-end nature of deep learning has made network framework selection and design the primary factor in sensing system implementations.

Table 10.

Automatic deep pattern-based sensing.

Year Reference Methodology Performance Base Signal Sensing Range Setting
2020 WiPose [77] CNN + LSTM 2.83 cm average error 3D velocity profile of CSI Single room 1000 Hz; three pairs of Tx-Rx; distributed deployed receiving Antennas
2020 WiSIA [78] cGAN Not applicable Power of CSI 2.1 m 1000 Hz; two pairs od Tx-Rx; receiving antennas orthogonal to each other
2021 Kang et al. [34] Adversarial learning and attention scheme 3–12.7% improvement DFS of CSI 2 m × 2 m Two pairs of Tx-Rx from Widar dataset
2022 GaitSense [27] CNN + LSTM + transfer learning + data augmentation 98% accuracy Gait-BVP of CSI 4.6 m × 4.4 m 1000 Hz; six pairs of Tx-Rx
2021 Ma et al. [46] CNN + reinforcement learning 97% average accuracy CSI amplitude 6.8 m × 4 m 100 Hz; single pair of Tx-Rx
2021 MCBAR [47] GAN and semi-supervised learning 90% average accuracy CSI amplitude 6.5 m × 6.3 m single pair of Tx-Rx
2021 WiFi-Sleep [73] Respiration and movement pattern extraction + CNN-BiLSTM 81.8% accuracy CSI ratio Close to the bed 200 Hz; single pair of Tx-Rx
2022 CAUTION [29] Few-shot learning 93.06 average accuracy CSI amplitude 5.2 m × 7.2 m Single pair of Tx-Rx
2022 Ding et al. [50] DCN + transfer learning 96.85% average accuracy CSI 6 m × 8 m 200 Hz; single pair of Tx-Rx
2022 EfficientFi [51] DNN 98% accuracy CSI amplitude 6.5 m × 5 m 500 Hz; single pair of Tx-Rx
2022 GoPose [79] 2D AOA spectrum + CNN + LSTM 93.2% for 5 users and 76.2% for 11 users CSI phase 4 m × 4 m 1000 Hz; four pairs of Tx-Rx; L-shaped receiving antennas
2022 ResFi [75] CNN-based classification 95% accuracy CSI amplitude 1 m 10 Hz; single pair of Tx-Rx
2022 TOSS [52] Meta learning + pseudo label strategy 82.69% average accuracy CSI Single room Single pair of Tx-Rx
2022 Widar 3.0 [39] BVP feature + CNN-RNN 92.7% in-domain and 82.6–92.4% cross-domain accuracy BVP of CSI 2 m × 2 m 1000 Hz; six pairs of Tx-Rx
2022 WiFine [40] data enhancement-based feature extraction + lightweight neural network 96.03% accuracy in 0.19 s CSI Single room Single pair of Tx-Rx
2022 Wi-PIGR [30] Spectrogram optimization + CNN + LSTM 93.5% for single user and 77.15% for 50 users CSI amplitude 5 m × 5 m 1000 Hz; two pairs of Tx-Rx
2023 Auto-Fi [31] Geometric self-supervised learning + few-shot calibration 86.83% for gesture; 79.61% for gait CSI amplitude Single room 100 Hz; single pair of Tx-Rx
2023 GaitFi [32] RCN + LSTM + feature fusion 94.2% accuracy CSI + video 2.1 m 800 Hz; single pair of Tx-Rx
2023 MetaFi++ [81] CNN + Transformer 97.3% for PCK@50 CSI + video Single room 1000 Hz; single pair of Tx-Rx
2023 FallDar [53] Scattering model + VAE generative model + DNN adversarial learning model 5.7% false alarm rate and 3.4% missed alarm rate ACF of CSI 3.6 m × 8.4 m 1000 Hz; single pair of Tx-Rx
2023 SHARP [54] Phase correction-based DFS extraction + Nerual network 95% average accuracy CSI 5 m × 6 m 173 Hz; single pair of Tx-Rx
2023 UniFi [41] DFS extraction + consistency-guided multi-view deep network + mutual information-based regularization 99% and 90–98% accuracy for in-domain and cross-domain recognition CSI ratio 2 m × 2 m Widar dataset
2023 WiTransformer [42] Transformer 86.16% accuracy BVP of CSI 2 m × 2 m Widar dataset
2024 AirFi [43] Data augmentation + adversarial learning +domain generalization 90% accuracy CSI amplitude 4 m × 4 m Single pair of Tx-Rx
2024 i-Sample [57] Intermediate sample generation + domain adversarial adaptation 10% accuracy gain CSI Single room Single pair of Tx-Rx
2024 MaskFi [58] Transformer-based encoder + Gate Recurrent Unit network 97.61% average accuracy CSI + video Single room 1000 Hz; Single pair of Tx-Rx
2024 MetaFormer [59] Transformer-based spatial-temporal feature extraction + match-based meta-learning approach Improved accuracy in various cross-domain scenarios CSI Single room SiFi, Widar, Wiar datasets
2024 PowerSkel [83] Knowledge distillation network based on collaborative learning and self-attention 96.27% for PCK@50 CSI + Kinect video Single room Three pairs of Tx-Rx
2024 SAT [60] Calibrated confidence-based adversarial sample selection + adversarial learning Improved accuracy and robustness CSI Single room Single pair of Tx-Rx
2024 SecureSense [61] Consistency-guided adversarial learning Robust performance under various attacks CSI amplitude 5 m × 6.5 m 1000 Hz; single pair of Tx-Rx
2024 Luo et al. [62] Transformer 98.78% accuracy CSI Single room UT-HAR dataset
2024 Wi-Diag [33] Independent component analysis-based blind source separation + CycleGAN 87.77% average accuracy CSI 7 m × 8 m 1000 Hz; single pair of Tx-Rx
2024 WiSMLF [63] High frequency energy-based sensing scheme selection + VGG/LSTM-based multi-level feature fusion 92% average accuracy CSI Single room 100 Hz; single pair of Tx-Rx
2024 Zhu et al. [25] ResNet18 95.57% average accuracy Amplified ACF of CSI 6 m × 6.5 m 1500 Hz; single pair of Tx-Rx

Apart from the above differences, we can obtain several additional findings from Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10. First, apart from the CSI amplitude and phase information, several new base signals, such as the BVP of CSI, ACF of CSI, and CSI ratio, have been used for alleviating the intrinsic errors of COTS WiFi devices [88]. Among these base signals, the CSI ratio is drawing more attention since it can not only remove the CSI offset, but it can also increase the sensing signal-to-noise rate (SNR) [89]. Second, some works have tried to combine a pattern-based scheme with model-based scheme to ensure the performance and reliability of complex sensing applications. Third, many systems have been developed for single human sensing under constrained deployment, i.e., single room sensing area with the LOS condition satisfied.

4. Challenges

Despite the above endeavors devoted to bringing Wi-Fi sensing from laboratory study to real-life applications, either by improving sensing granularity or exploring application scenarios, most of the existing works still face great practical challenges. Specifically, making Wi-Fi sensing system readily available for wide real-world deployment, easily adaptable to different environments, and with enough sensing coverage is of vital importance. This section presents two key challenges faced in existing works, i.e., the domain dependent issue and the sensing range limitation, and it discusses related potential solution explorations.

Domain dependent issue. As the superposition result of multi-path signals, Wi-Fi is highly sensitive to various factors, such as locations, orientations, targets, and environments. This is also known as the domain-dependence problem [15,18,86]. For example, the same human activity will lead to quite different CSI variations if the location or orientation of the target changes, as revealed by the Fresnel zone model. Moreover, different sensing environments and device settings will make this inconsistent phenomenon even worse. A sensing system lacking resilience to domain variations is in fact of little practical use for ubiquitous sensing. Thus, in order to make Wi-Fi sensing reusable and robust among different settings, researchers have explored various methods, as summarized in Table 11. Since training effort accounts for a great part of the system deployment cost, Table 11 classifies the related works into three categories, i.e., training-free, training-once, and training + Calibration/Retrained. As seen in Table 11, the training-free scheme is mainly used for simple presence detection tasks [23,24], where a motion-induced threshold is predetermined without training. In addition, for the training-once scheme, the domain-independent feature extraction is the most studied, and it is mainly used along model-based sensing, as listed in Table 8 due to clear interpretability. Moreover, with the increasing complexity of sensing tasks and environments, system recalibration would become inevitable, promoting researchers to reduce the system retraining cost, e.g., utilizing data augmentation, transfer learning, and few-shot learning, as shown in Table 11. It can also be observed that domain-independent feature extraction can be used alone or further integrated with other retraining algorithms. Drawn from the above discussions, for these complex applications, it is expected that combining the strengths of model-based and auto deep learning model-based methods can enable a more general and robust Wi-Fi sensing realization.

Table 11.

Cross-domain Wi-Fi sensing.

Training Cost Cross-Domain Scheme Related Work
Training-free Domain-independent feature extraction [23,24]
Training-once Domain-independent feature extraction [25,26,27,28,30,34,35,36,37,38,39,41,42,44,49,53,54,64,65,66,72]
Training + Calibration/Retrained Generative adversarial network [33,47,53,61]
Transfer learning [27,31,34,43,50,57,60]
Few-shot learning [29,31,43,52]
Data augmentation [27,43,57]
CNN +LSTM/GRU/Transformer [25,30,32,39,41,42,46,58,59,62,81]

Sensing range limitation. As illustrated in the tables of last section, the existing sensing range is usually just 6–8 m within a single room, while the communication range of Wi-Fi can reach tens of meters, greatly hindering real-world applications. The short sensing range is mainly because Wi-Fi sensing relies on target-induced reflection signal variation, which is much weaker compared to direct LOS signal and contains intrinsic hardware noise. To be more specific, due to hardware imperfections and clock synchronization errors, the raw CSI amplitude contains high impulse and burst noise, while the raw randomly corrupted CSI phase is even more unusable in practice. To deal with this limitation, some researchers proposed employing a new base signal derived from the raw CSI, namely the CSI ratio as seen in Table 8 and Table 9. Defined as the quotient of CSI readings between two receiver antennas, the CSI ratio can remove the amplitude noise and phase noise effectively. More specifically, since different antennas on the same receiver share the same RF chain and clock, the division operation can cancel out most of the noise, gaining a more ideal amplitude and a phase signal with a high signal-to-noise ratio (SNR). The higher SNR and phase usability of the CSI ratio serve as the key enablers for the longer sensing range and higher sensing accuracy. FarSense [90] first increased fine-grained sensing range to 8 m using the CSI ratio signal, while Zeng et al. [91] and DiverSense [92] further boosted the sensing range to 18 m and 40 m by fully utilizing the spatial and frequency diversity. In addition to constructing a new base signal, Wang et al. [93] studied the effect of device placement on sensing SNR and doubly expanded the sensing range by properly placing the transmitter and receiver. Overall, sensing range enlargement is pivoted to large-scale sensing applications and is still in its infancy, requiring further exploration and validation in complex real-world scenario deployments.

5. Future Research Trend Discussion

Despite the great effort spent on Wi-Fi sensing over the past years, there still exists a great gap for its pervasive real-life application. Based on the detailed analysis above, we point out three critical barriers that require further research in this section.

Table 12.

CSI extraction tools.

Year CSI Extraction Tool IEEE Standard Related Work
2011 802.11n CSI Tool [17] 802.11n [27,28,30,33,35,36,37,38,39,44,46,48,49,50,52,53,56,57,60,63,64,65,66,71,73,77,78,79,80,82,84]
2015 Atheros CSI Tool [94] 802.11n [29,31,32,47,51,58,61,72,81,94]
2019 Nexmon CSI [95] 802.11 ac [40,54,74,75,95]
2020 ESP32 CSI Tool [96,97] Any computer, smartphone or even standalone [83,96,97]
2021 AX-CSI [98] 802.11 ax [98]
2022 PicoScenes [99] 802.11 a/g/n/ac/ax [70,99]

Table 13.

Wi-Fi sensing datasets.

Year Dataset Description Tool Related Work
2017 UT-HAR [100] Activity data 802.11n CSI Tool [31,46,62]
2018 SignFi [101] Sign data 802.11n CSI Tool [40,59]
2018 FallDeFi [102] Fall data 802.11n CSI Tool [46,53]
2019 WiAR [103] Activity and gesture data 802.11n CSI Tool [59]
2019 Widar [104] Gesture data 802.11n CSI Tool [31,34,39,41,42,43,59]
2021 OneFi [105] Gesture data 802.11n CSI Tool [105]
2023 MM-Fi [106] Multi-modal dataset Atheros CSI Tool [58]
2023 NTU-Fi [107] Activity and gait data Atheros CSI Tool [62]
2023 SHARP [54] Activity data Nexmon CSI [54]
2023 Cominelli [108] Activity data AX-CSI [108]
2023 WiTraj [65] Trajectory data 802.11n CSI Tool [65]

Sensing assessment standardization. One key issue is the lack of a standard performance evaluation of the various Wi-Fi sensing systems. Unlike the widely accepted standard evaluation criterion in the computer vision domain, there is still a lack of an effective and consistent testing platform in Wi-Fi sensing. Specifically, the deficiency exists in two aspects, i.e., CSI extraction tool diversity and evaluation dataset scarcity. The diversity of CSI extraction tools is shown in Table 12, with Intel 5300 NIC-based 802.11n CSI Tool being the most popular one used. However, sensing techniques developed with old 802.11n protocol have not explored the innovations of newer standards and may even fail when used on new-generation Wi-Fi cards [108,109]. In addition, as illustrated in Table 13, although there have been some publicly released datasets, none of them have been widely used. Existing works mostly adopt self-collected datasets, collected in different scenarios with different tools, hindering the comparability and replicability of research outcomes. To build comprehensive datasets without labor-intensive and time-consuming efforts, researchers have studied radio signal synthesis [110,111] and physical data augmentation [112], providing promising solutions to the data scarcity problem. We believe a more unified CSI extraction tool compatible with the new 802.11 standard and a set of standard datasets for a benchmark comparison should be indispensable for the further research cooperation and development of Wi-Fi sensing.

Sensing and communication balance. As illustrated in Table 14, most sensing systems require a high sampling rate for reliable performance, which interferes with regular Wi-Fi communication. To be more specific, the data throughput undergoes great drop when the sampling rate for sensing is higher than 50 Hz [66]. SenCom [113] managed to extract CSI from general communication packets and obtained evenly sampled and sufficient CSI data with a detailed signal processing technique. While appealing, SenCom is not yet applicable for COTS clients. Thus, the ways of enabling Wi-Fi sensing while maintaining communication capability, i.e., achieving sensing and communication balance, remain an open problem in the current ISAC area.

Sensing generalization and reliability. As noted in Table 12, raw CSI reading is still only accessible with limited hardware; some researchers resorted to sensing with other Wi-Fi signals. For instance, since the beamforming feedback matrix (BFM) is readily available with all new-generation MU-MIMO-enabled Wi-Fi cards, researchers have explored generalized Wi-Fi sensing using BFM [114,115]. In addition, to improve the reliability of sensing, multi-modal sensing, which integrates Wi-Fi and other sensing modalities, e.g., video [32,52,81,116] and received signal strength indicator (RSSI) [117], are worth further studying.

Table 14.

Sampling rate of recent works.

Sampling Rate Related Work
≤100 Hz [23,24,31,45,46,55,63,66,72,74,75,83]
100 Hz–500 Hz [35,36,37,44,48,50,51,54,65,71,73,82,84]
>500 Hz [25,26,27,28,30,32,33,38,39,49,53,56,58,64,77,78,79,81]

Apart from the above discussion, the physical challenges of the existing Wi-Fi infrastructure should also be noticed, which will greatly determine the possible sensing limit of Wi-Fi sensing. First, due to hardware and network design, clock asynchronism between Wi-Fi transmitter and receiver is a severe issue in an ISAC system. It introduces a time-varying random phase offset in raw CSI, making reliable feature extraction difficult. Second, except for target influence, dynamic parameter adjustments of the network card during transmission also affect the CSI measurement, which is highly dependent on the hardware design. Third, large-scale Wi-Fi sensing needs to obtain CSI from multiple distributed receivers. The ways of enabling CSI estimation and alignment over multiple devices are a challenging problem. Currently, there is no universal solution to the above challenges, requiring cooperative efforts from application researchers, chip manufactures, and communication equipment vendors.

6. Conclusions

Owing to the active participation from numerous researchers, notable advances have been made in Wi-Fi sensing techniques in recent years. In an effort to gain insight into future trends, this paper reviews major achievements over the last 5 years and carries out an in-depth analysis of various methods, including limitations and practical challenges faced in existing systems. Moreover, to realize massive real-life applications, this paper highlights three imperative and promising future directions which are as follows: sensing assessment standardization, sensing and communication balance, and sensing generalization and reliability. We hope this review can help people better understand the progress and problems within the current Wi-Fi sensing research field, inspiring more amazing ideas for the upcoming ubiquitous ISAC.

Appendix A

Table A1 summaries all references mentioned in Section 3, pointing out corresponding application and methodology categories, and can direct the interested reader to related subsections for a more detailed description.

Table A1.

Summary of references in Section 3.

Year Reference Application Methodology Related Subsections
2022 WiCPD [23] In-car child presence detection Feature-based motion, stationary and transition target detector Table 1 in Section 3.1;
Table 9 in Section 3.2
2023 Hu et al. [24] Proximity detection Sub-carrier correlation and covariance feature extraction Table 1 in Section 3.1;
Table 9 in Section 3.2
2024 Zhu et al. [25] Human and non-human differentiation ResNet18 Table 1 in Section 3.1;
Table 10 in Section 3.2
2024 WI-MOID [26] Edge device-based human and non-human differentiation Physical and statistical pattern extraction + SVM + state machine Table 1 in Section 3.1;
Table 9 in Section 3.2
2021 GaitSense [27] Gait-based human identification CNN + LSTM + transfer learning + data augmentation Table 2 in Section 3.1;
Table 10 in Section 3.2
2021 GaitWay [28] Gait speed estimation Scattering model Table 2 in Section 3.1;
Table 8 in Section 3.2
2022 CAUTION [29] Gait-based human authentication Few-shot learning Table 2 in Section 3.1;
Table 10 in Section 3.2
2022 Wi-PIGR [30] Gait recognition Spectrogram optimization + CNN + LSTM Table 2 in Section 3.1;
Table 10 in Section 3.2
2023 Auto-Fi [31] Gesture and gait recognition Geometric self-supervised learning + few-shot calibration Table 2 in Section 3.1;
Table 10 in Section 3.2
2023 GaitFi [32] Gait recognition RCN + LSTM + feature fusion Table 2 in Section 3.1;
Table 10 in Section 3.2
2024 Wi-Diag [33] Multi-subject abnormal gait diagnosis Independent component analysis-based blind source separation + CycleGAN Table 2 in Section 3.1;
Table 10 in Section 3.2
2021 Kang et al. [34] Gesture recognition Adversarial learning and attention scheme Table 3 in Section 3.1;
Table 10 in Section 3.2
2021 WiGesture [35] Gesture recognition MNP feature extraction Table 3 in Section 3.1;
Table 9 in Section 3.2
2022 HandGest [36] Handwriting recognition Hand-centric feature extraction, i.e., DPV and MRV Table 3 in Section 3.1;
Table 9 in Section 3.2
2022 DPSense-WiGesture [37] Gesture recognition Signal segmentation + sensing quality-based signal processing Table 3 in Section 3.1;
Table 9 in Section 3.2
2022 Niu et al. [38] Gesture recognition Position-independent feature extraction, i.e., movement fragment and relative motion direction change Table 3 in Section 3.1;
Table 9 in Section 3.2
2022 Widar 3.0 [39] Cross-domain gesture recognition BVP feature + CNN-RNN Table 3 in Section 3.1;
Table 10 in Section 3.2
2022 WiFine [40] Gesture recognition Data enhancement-based feature extraction + lightweight neural network Table 3 in Section 3.1;
Table 10 in Section 3.2
2023 UniFi [41] Gesture recognition DFS extraction + consistency-guided multi-view deep network + mutual information-based regularization Table 3 in Section 3.1;
Table 10 in Section 3.2
2023 WiTransformer [42] Gesture recognition Transformer Table 3 in Section 3.1;
Table 10 in Section 3.2
2024 AirFi [43] Gesture recognition Data augmentation + adversarial learning + domain generalization Table 3 in Section 3.1;
Table 10 in Section 3.2
2024 WiCGesture [44] Continuous gesture recognition Meta motion-based signal segmentation and back-tracking searching-based identification Table 3 in Section 3.1;
Table 9 in Section 3.2
2020 Wang et al. [45] People counting and recognition Statistical pattern analysis Table 4 in Section 3.1;
Table 9 in Section 3.2
2021 Ma et al. [46] Activity recognition CNN + reinforcement learning Table 4 in Section 3.1;
Table 10 in Section 3.2
2021 MCBAR [47] Activity recognition GAN and semi-supervised learning Table 4 in Section 3.1;
Table 10 in Section 3.2
2021 WiMonitor [48] Location and activity monitoring Doppler frequency and activity intensity pattern extraction Table 4 in Section 3.1;
Table 9 in Section 3.2
2022 DeFall [49] Fall detection Scattering model Table 4 in Section 3.1;
Table 8 in Section 3.2
2022 Ding et al. [50] Activity recognition DCN + transfer learning Table 4 in Section 3.1;
Table 10 in Section 3.2
2022 EfficientFi [51] Activity recognition DNN Table 4 in Section 3.1;
Table 10 in Section 3.2
2022 TOSS [52] Activity recognition Meta learning + pseudo label strategy Table 4 in Section 3.1;
Table 10 in Section 3.2
2023 FallDar [53] Fall detection Scattering model + VAE generative model + DNN adversarial learning model Table 4 in Section 3.1;
Table 10 in Section 3.2
2023 SHARP [54] Activity recognition Phase correction-based DFS extraction + Nerual network Table 4 in Section 3.1;
Table 10 in Section 3.2
2023 Liu et al. [55] Moving receiver-based activity recognition Dynamic Fresnel zone model Table 4 in Section 3.1;
Table 8 in Section 3.2
2023 WiCross [56] Target passing detection Diffraction model-based phase pattern extraction Table 4 in Section 3.1;
Table 8 in Section 3.2
2024 i-Sample [57] Activity recognition Intermediate sample generation + domain adversarial adaptation Table 4 in Section 3.1;
Table 10 in Section 3.2
2024 MaskFi [58] Activity recognition Transformer-based encoder + Gate Recurrent Unit network Table 4 in Section 3.1;
Table 10 in Section 3.2
2024 MetaFormer [59] Activity recognition Transformer-based spatial-temporal feature extraction + match-based meta-learning approach Table 4 in Section 3.1;
Table 10 in Section 3.2
2024 SAT [60] Activity recognition Calibrated confidence-based adversarial sample selection + adversarial learning Table 4 in Section 3.1;
Table 10 in Section 3.2
2024 SecureSense [61] Activity recognition under adversarial attack Consistency-guided adversarial learning Table 4 in Section 3.1;
Table 10 in Section 3.2
2024 Luo et al. [62] Activity recognition Transformer Table 4 in Section 3.1;
Table 10 in Section 3.2
2024 WiSMLF [63] Activity recognition High-frequency energy-based sensing scheme selection + VGG/LSTM-based multi-level feature fusion Table 4 in Section 3.1;
Table 10 in Section 3.2
2022 Niu et al. [64] Velocity estimation-based tracing DFS-based velocity estimation + receiver selection Table 5 in Section 3.1;
Table 9 in Section 3.2
2023 WiTraj [65] Human walking tracking DFS extraction + multi-view trajectory estimation + motion detection Table 5 in Section 3.1;
Table 9 in Section 3.2
2024 FewSense [66] Tracking TD-CSI-based Doppler speed estimation Table 5 in Section 3.1;
Table 9 in Section 3.2
2023 Zhang et al. [67] Multi-person localization 2D antenna array-based signal parameter estimation Table 5 in Section 3.1;
Table 9 in Section 3.2
2024 Zhang et al. [68] Passive localization Rigid-body coordinate transformation-based absolute pose and relative motion estimation Table 5 in Section 3.1;
Table 9 in Section 3.2
2022 Fan et al. [69] Moving direction estimation Intelligent reflecting surface construction Table 5 in Section 3.1;
Table 9 in Section 3.2
2022 Wi-Drone [70] Tracking-based indoor drone flight control Intelligent reflecting surface construction Table 5 in Section 3.1;
Table 9 in Section 3.2
2020 MultiSense [71] Multi-person respiration sensing ICA-based BSS Table 6 in Section 3.1;
Table 9 in Section 3.2
2021 SMARS [72] Breath estimation and sleep stage recognition Scattering model Table 6 in Section 3.1;
Table 8 in Section 3.2;
2021 WiFi-Sleep [73] Sleep stage monitoring Respiration and movement pattern extraction + CNN-BiLSTM Table 6 in Section 3.1;
Table 10 in Section 3.2
2021 WiPhone [74] Respiration monitoring Ambient reflection-based pattern extraction Table 6 in Section 3.1;
Table 9 in Section 3.2
2022 ResFi [75] Respiration detection CNN-based classification Table 6 in Section 3.1;
Table 10 in Section 3.2
2024 Xie et al. [76] Respiration sensing with interfering individual Respiratory energy-based interference detection and convex optimization-based beam control Table 6 in Section 3.1;
Table 9 in Section 3.2
2020 WiPose [77] Pose construction CNN + LSTM Table 7 in Section 3.1;
Table 10 in Section 3.2
2020 WiSIA [78] Target imaging cGAN Table 7 in Section 3.1;
Table 10 in Section 3.2
2022 GoPose [79] 3D human pose estimation 2D AOA spectrum + CNN + LSTM Table 7 in Section 3.1;
Table 10 in Section 3.2
2022 Wiffract [80] Still object imaging Keller’s Geometrical Theory of Diffraction Table 7 in Section 3.1;
Table 8 in Section 3.2
2023 MetaFi++ [81] Pose estimation CNN + Transformer Table 7 in Section 3.1;
Table 10 in Section 3.2
2023 WiMeasure [82] Object size measurement Diffraction model Table 7 in Section 3.1;
Table 8 in Section 3.2
2024 PowerSkel [83] Pose estimation Knowledge distillation network based on collaborative learning and self-attention Table 7 in Section 3.1;
Table 10 in Section 3.2
2024 WiProfile [84] 2D target Profiling Diffraction effect-based profiling + inverse Fresnel transform Table 7 in Section 3.1;
Table 8 in Section 3.2

Author Contributions

Conceptualization, H.Z.; writing—original draft preparation, H.Z., E.D. and M.X.; discussion and supervision, H.L. and F.W. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

Funding Statement

This work was supported in part by the Young Scientists Fund of the National Natural Science Foundation of China, under grants 61902237 and 52205597, and the Key Project of Science and Technology Commission of Shanghai Municipality, under grant 22DZ1100803.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Liu A., Huang Z., Li M., Wan Y., Li W., Han T.X., Liu C., Du R., Tan D.K.P., Lu J., et al. A Survey on Fundamental Limits of Integrated Sensing and Communication. IEEE Commun. Surv. Tutor. 2022;24:994–1034. doi: 10.1109/COMST.2022.3149272. [DOI] [Google Scholar]
  • 2.Liu F., Cui Y., Masouros C., Xu J., Han T.X., Eldar Y.C., Buzzi S. Integrated Sensing and Communications: Toward Dual-Functional Wireless Networks for 6G and Beyond. IEEE J. Sel. Areas Commun. 2022;40:1728–1767. doi: 10.1109/JSAC.2022.3156632. [DOI] [Google Scholar]
  • 3.Meneghello F., Chen C., Cordeiro C., Restuccia F. Toward Integrated Sensing and Communications in IEEE 802.11bf Wi-Fi Networks. IEEE Commun. Mag. 2023;61:128–133. doi: 10.1109/MCOM.001.2200806. [DOI] [Google Scholar]
  • 4.Wu C., Wang B., Au O., Liu K. Wi-Fi Can Do More: Toward Ubiquitous Wireless Sensing. IEEE Commun. Stand. Mag. 2022;6:42–49. doi: 10.1109/MCOMSTD.0001.2100111. [DOI] [Google Scholar]
  • 5.Li X., Cui Y., Zhang J., Liu F., Zhang D., Hanzo L. Integrated Human Activity Sensing and Communications. IEEE Commun. Mag. 2023;61:90–96. doi: 10.1109/MCOM.002.2200391. [DOI] [Google Scholar]
  • 6.Yang Z., Zhou Z., Liu Y. From RSSI to CSI: Indoor localization via channel response. ACM Comput. Surv. 2013;46:1–32. doi: 10.1145/2543581.2543592. [DOI] [Google Scholar]
  • 7.Zhang F., Wu C., Wang B., Lai H., Han Y., Ray Liu K. WiDetect: Robust Motion Detection with a Statistical Electromagnetic Model. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019;3:1–24. doi: 10.1145/3351280. [DOI] [Google Scholar]
  • 8.Zhang F., Niu K., Xiong J., Jin B., Gu T., Jiang Y., Zhang D. Towards a Diffraction-based Sensing Approach on Human Activity Recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019;3:1–25. doi: 10.1145/3314420. [DOI] [Google Scholar]
  • 9.Gong W., Liu J. SiFi: Pushing the Limit of Time-Based WiFi Localization Using a Single Commodity Access Point. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;2:1–21. doi: 10.1145/3191742. [DOI] [Google Scholar]
  • 10.Zhang D., Wang H., Wu D. Toward Centimeter-Scale Human Activity Sensing with Wi-Fi Signals. Computer. 2017;50:48–57. doi: 10.1109/MC.2017.7. [DOI] [Google Scholar]
  • 11.Wang Z., Jiang K., Hou Y., Dou W., Zhang C., Huang Z., Guo Y. A Survey on Human Behavior Recognition Using Channel State Information. IEEE Access. 2019;7:155986–156024. doi: 10.1109/ACCESS.2019.2949123. [DOI] [Google Scholar]
  • 12.Ma Y., Zhou G., Wang S. WiFi Sensing with Channel State Information: A Survey. ACM Comput. Surv. 2019;52:1–36. doi: 10.1145/3310194. [DOI] [Google Scholar]
  • 13.Tan S., Ren Y., Yang J., Chen Y. Commodity WiFi Sensing in Ten Years: Status, Challenges, and Opportunities. IEEE Internet Things J. 2022;9:17832–17843. doi: 10.1109/JIOT.2022.3164569. [DOI] [Google Scholar]
  • 14.Xiao J., Li H., Wu M., Jin H., Jamal Deen M., Cao J. A Survey on Wireless Device-free Human Sensing: Application Scenarios, Current Solutions, and Open Issues. ACM Comput. Surv. 2022;55:1–35. doi: 10.1145/3530682. [DOI] [Google Scholar]
  • 15.Chen C., Zhou G., Lin Y. Cross-Domain WiFi Sensing with Channel State Information: A Survey. ACM Comput. Surv. 2023;55:1–37. doi: 10.1145/3570325. [DOI] [Google Scholar]
  • 16.Halperin D., Hu W., Sheth A., Wetherall D. Predictable 802.11 packet delivery from wireless channel measurements. SIGCOMM Comput. Commun. Rev. 2010;40:159–170. doi: 10.1145/1851275.1851203. [DOI] [Google Scholar]
  • 17.Halperin D., Hu W., Sheth A., Wetherall D. Tool release: Gathering 802.11n traces with channel state information. SIGCOMM Comput. Commun. Rev. 2011;41:53. doi: 10.1145/1925861.1925870. [DOI] [Google Scholar]
  • 18.Wang H., Zhang D., Ma J., Wang Y., Wang Y., Wu D., Gu T., Xie B. Human respiration detection with commodity wifi devices: Do user location and body orientation matter?; Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp ‘16); Heidelberg, Germany. 12–16 September 2016; New York, NY, USA: Association for Computing Machinery; 2016. pp. 25–36. [Google Scholar]
  • 19.Wu D., Zhang D., Xu C., Wang H., Li X. Device-Free WiFi Human Sensing: From Pattern-Based to Model-Based Approaches. IEEE Commun. Mag. 2017;55:91–97. doi: 10.1109/MCOM.2017.1700143. [DOI] [Google Scholar]
  • 20.Zhang F., Zhang D., Xiong J., Wang H., Niu K., Jin B., Wang Y. From Fresnel Diffraction Model to Fine-grained Human Respiration Sensing with Commodity Wi-Fi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;2:1–23. doi: 10.1145/3191785. [DOI] [Google Scholar]
  • 21.Yang Z., Zhang Y., Chi G., Zhang G. Hands-on wireless sensing with Wi-Fi: A tutorial. arXiv. 20222206.09532 [Google Scholar]
  • 22.Zhang F., Chen C., Wang B., Liu K.J.R. WiSpeed: A Statistical Electromagnetic Approach for Device-Free Indoor Speed Estimation. IEEE Internet Things J. 2018;5:2163–2177. doi: 10.1109/JIOT.2018.2826227. [DOI] [Google Scholar]
  • 23.Zeng X., Wang B., Wu C., Regani S.D., Liu K.J.R. WiCPD: Wireless Child Presence Detection System for Smart Cars. IEEE Internet Things J. 2022;9:24866–24881. doi: 10.1109/JIOT.2022.3194873. [DOI] [Google Scholar]
  • 24.Hu Y., Ozturk M.Z., Wang B., Wu C., Zhang F., Liu K.J.R. Robust Passive Proximity Detection Using Wi-Fi. IEEE Internet Things J. 2023;10:6221–6234. doi: 10.1109/JIOT.2022.3224701. [DOI] [Google Scholar]
  • 25.Zhu G., Wang B., Gao W., Hu Y., Wu C., Liu K.J.R. WiFi-Based Robust Human and Non-human Motion Recognition With Deep Learning; Proceedings of the 2024 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops); Biarritz, France. 11–15 March 2024; pp. 769–774. [Google Scholar]
  • 26.Zhu G., Hu Y., Wang B., Wu C., Zeng X., Liu K.J.R. Wi-MoID: Human and Nonhuman Motion Discrimination Using WiFi With Edge Computing. IEEE Internet Things J. 2024;11:13900–13912. doi: 10.1109/JIOT.2023.3339544. [DOI] [Google Scholar]
  • 27.Zhang Y., Zheng Y., Zhang G., Qian K., Qian C., Yang Z. GaitSense: Towards Ubiquitous Gait-Based Human Identification with Wi-Fi. ACM Trans. Sens. Netw. 2021;18:1–24. doi: 10.1145/3466638. [DOI] [Google Scholar]
  • 28.Wu C., Zhang F., Hu Y., Liu K.J.R. GaitWay: Monitoring and Recognizing Gait Speed Through the Walls. IEEE Trans. Mob. Comput. 2021;20:2186–2199. doi: 10.1109/TMC.2020.2975158. [DOI] [Google Scholar]
  • 29.Wang D., Yang J., Cui W., Xie L., Sun S. CAUTION: A Robust WiFi-Based Human Authentication System via Few-Shot Open-Set Recognition. IEEE Internet Things J. 2022;9:17323–17333. doi: 10.1109/JIOT.2022.3156099. [DOI] [Google Scholar]
  • 30.Zhang L., Wang C., Zhang D. Wi-PIGR: Path Independent Gait Recognition With Commodity Wi-Fi. IEEE Trans. Mob. Comput. 2022;21:3414–3427. doi: 10.1109/TMC.2021.3052314. [DOI] [Google Scholar]
  • 31.Yang J., Chen X., Zou H., Wang D., Xie L. AutoFi: Toward Automatic Wi-Fi Human Sensing via Geometric Self-Supervised Learning. IEEE Internet Things J. 2023;10:7416–7425. doi: 10.1109/JIOT.2022.3228820. [DOI] [Google Scholar]
  • 32.Deng L., Yang J., Yuan S., Zou H., Lu C.X., Xie L. GaitFi: Robust Device-Free Human Identification via WiFi and Vision Multimodal Learning. IEEE Internet Things J. 2023;10:625–636. doi: 10.1109/JIOT.2022.3203559. [DOI] [Google Scholar]
  • 33.Zhang L., Ma Y., Fan X., Fan X., Zhang Y., Chen Z., Chen X., Zhang D. Wi-Diag: Robust Multisubject Abnormal Gait Diagnosis With Commodity Wi-Fi. IEEE Internet Things J. 2024;11:4362–4376. doi: 10.1109/JIOT.2023.3301908. [DOI] [Google Scholar]
  • 34.Kang H., Zhang Q., Huang Q. Context-Aware Wireless-Based Cross-Domain Gesture Recognition. IEEE Internet Things J. 2021;8:13503–13515. doi: 10.1109/JIOT.2021.3064890. [DOI] [Google Scholar]
  • 35.Gao R., Zhang M., Zhang J., Li Y., Yi E., Wu D., Wang L., Zhang D. Towards Position-Independent Sensing for Gesture Recognition with Wi-Fi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021;5:1–28. doi: 10.1145/3463504. [DOI] [Google Scholar]
  • 36.Zhang J., Li Y., Xiong H., Dou D., Miao C., Zhang D. HandGest: Hierarchical Sensing for Robust-in-the-Air Handwriting Recognition With Commodity WiFi Devices. IEEE Internet Things J. 2022;9:19529–19544. doi: 10.1109/JIOT.2022.3170157. [DOI] [Google Scholar]
  • 37.Gao R., Li W., Xie Y., Yi E., Wang L., Wu D., Zhang D. Towards Robust Gesture Recognition by Characterizing the Sensing Quality of WiFi Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–26. doi: 10.1145/3517241. [DOI] [Google Scholar]
  • 38.Niu K., Zhang F., Wang X., Lv Q., Luo H., Zhang D. Understanding WiFi Signal Frequency Features for Position-Independent Gesture Sensing. IEEE Trans. Mob. Comput. 2022;21:4156–4171. doi: 10.1109/TMC.2021.3063135. [DOI] [Google Scholar]
  • 39.Zheng Y., Zheng Y., Qian K., Zhang G., Liu Y., Wu C., Yang Z. Widar3.0: Zero-Effort Cross-Domain Gesture Recognition with Wi-Fi. IEEE Trans. Pattern Anal. Mach. Intell. 2022;44:8671–8688. doi: 10.1109/TPAMI.2021.3105387. [DOI] [PubMed] [Google Scholar]
  • 40.Xing T., Yang Q., Jiang Z., Fu X., Wang J., Wu C.Q., Chen X. WiFine: Real-Time Gesture Recognition Using Wi-Fi with Edge Intelligence. ACM Trans. Sens. Netw. 2022;19:1–24. doi: 10.1145/3532094. [DOI] [Google Scholar]
  • 41.Liu Y., Yu A., Wang L., Guo B., Li Y., Yi E., Zhang D. UniFi: A Unified Framework for Generalizable Gesture Recognition with Wi-Fi Signals Using Consistency-guided Multi-View Networks. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;7:1–29. doi: 10.1145/3631429. [DOI] [Google Scholar]
  • 42.Yang M., Zhu H., Zhu R., Wu F., Yin L., Yang Y. WiTransformer: A Novel Robust Gesture Recognition Sensing Model with WiFi. Sensors. 2023;23:2612. doi: 10.3390/s23052612. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Wang D., Yang J., Cui W., Xie L., Sun S. AirFi: Empowering WiFi-Based Passive Human Gesture Recognition to Unseen Environment via Domain Generalization. IEEE Trans. Mob. Comput. 2024;23:1156–1168. doi: 10.1109/TMC.2022.3230665. [DOI] [Google Scholar]
  • 44.Gao R., Li W., Liu J., Dai S., Zhang M., Wang L., Zhang D. WiCGesture: Meta-Motion-Based Continuous Gesture Recognition With Wi-Fi. IEEE Internet Things J. 2024;11:15087–15099. doi: 10.1109/JIOT.2023.3343875. [DOI] [Google Scholar]
  • 45.Wang F., Zhang F., Wu C., Wang B., Liu K.J.R. Respiration Tracking for People Counting and Recognition. IEEE Internet Things J. 2020;7:5233–5245. doi: 10.1109/JIOT.2020.2977254. [DOI] [Google Scholar]
  • 46.Ma Y., Arshad S., Muniraju S., Torkildson E., Rantala E., Doppler K., Zhou G. Location- and Person-Independent Activity Recognition with WiFi, Deep Neural Networks, and Reinforcement Learning. ACM Trans. Internet Things. 2021;2:1–25. doi: 10.1145/3424739. [DOI] [Google Scholar]
  • 47.Wang D., Yang J., Cui W., Xie L., Sun S. Multimodal CSI-Based Human Activity Recognition Using GANs. IEEE Internet Things J. 2021;8:17345–17355. doi: 10.1109/JIOT.2021.3080401. [DOI] [Google Scholar]
  • 48.Niu X., Li S., Zhang Y., Liu Z., Wu D., Shah R.C., Tanriover C., Lu H., Zhang D. WiMonitor: Continuous Long-Term Human Vitality Monitoring Using Commodity Wi-Fi Devices. Sensors. 2021;21:751. doi: 10.3390/s21030751. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Hu Y., Zhang F., Wu C., Wang B., Liu K.J.R. DeFall: Environment-Independent Passive Fall Detection Using WiFi. IEEE Internet Things J. 2022;9:8515–8530. doi: 10.1109/JIOT.2021.3116136. [DOI] [Google Scholar]
  • 50.Ding X., Hu C., Xie W., Zhong Y., Yang J., Jiang T. Device-Free Multi-Location Human Activity Recognition Using Deep Complex Network. Sensors. 2022;22:6178. doi: 10.3390/s22166178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Yang J., Chen X., Zou H., Wang D., Xu Q., Xie L. EfficientFi: Toward Large-Scale Lightweight WiFi Sensing via CSI Compression. IEEE Internet Things J. 2022;9:13086–13095. doi: 10.1109/JIOT.2021.3139958. [DOI] [Google Scholar]
  • 52.Zhou Z., Wang F., Yu J., Ren J., Wang Z., Gong W. Target-oriented Semi-supervised Domain Adaptation for WiFi-based HAR; Proceedings of the IEEE INFOCOM 2022-IEEE Conference on Computer Communications; London, UK. 2–5 May 2022; pp. 420–429. [Google Scholar]
  • 53.Yang Z., Zhang Y., Zhang Q. Rethinking Fall Detection With Wi-Fi. IEEE Trans. Mob. Comput. 2023;22:6126–6143. doi: 10.1109/TMC.2022.3188779. [DOI] [Google Scholar]
  • 54.Meneghello F., Garlisi D., Di Fabbro N., Tinnirello I., Rossi M. SHARP: Environment and Person Independent Activity Recognition With Commodity IEEE 802.11 Access Points. IEEE Trans. Mob. Comput. 2023;22:6160–6175. doi: 10.1109/TMC.2022.3185681. [DOI] [Google Scholar]
  • 55.Liu J., Li W., Gu T., Gao R., Chen B., Zhang F., Wu D., Zhang D. Towards a Dynamic Fresnel Zone Model to WiFi-based Human Activity Recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2023;7:1–24. doi: 10.1145/3596270. [DOI] [Google Scholar]
  • 56.Shi W., Wang X., Niu K., Wang L., Zhang D. WiCross: I Can Know When You Cross Using COTS WiFi Devices; Proceedings of the Adjunct Proceedings of the 2023 ACM International Joint Conference on Pervasive and Ubiquitous Computing & the 2023 ACM International Symposium on Wearable Computing (UbiComp/ISWC ‘23 Adjunct); Cancun, Mexico. 8–12 October 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 133–136. [Google Scholar]
  • 57.Zhou Z., Wang F., Gong W. I-Sample: Augment Domain Adversarial Adaptation Models for WiFi-based HAR. ACM Trans. Sens. Netw. 2024;20:1–20. doi: 10.1145/3616494. [DOI] [Google Scholar]
  • 58.Yang J., Tang S., Xu Y., Zhou Y., Xie L. MaskFi: Unsupervised Learning of WiFi and Vision Representations for Multimodal Human Activity Recognition. arXiv. 20242402.19258 [Google Scholar]
  • 59.Sheng B., Han R., Xiao F., Guo Z., Gui L. MetaFormer: Domain-Adaptive WiFi Sensing with Only One Labelled Target Sample. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;8:1–27. doi: 10.1145/3643550. [DOI] [Google Scholar]
  • 60.Pan Y., Zhou Z., Gong W., Fang Y. SAT: A Selective Adversarial Training Approach for WiFi-based Human Activity Recognition. IEEE Trans. Mob. Comput. 2024;23:12706–12716. doi: 10.1109/TMC.2024.3420405. [DOI] [Google Scholar]
  • 61.Yang J., Zou H., Xie L. SecureSense: Defending Adversarial Attack for Secure Device-Free Human Activity Recognition. IEEE Trans. Mob. Comput. 2024;23:823–834. doi: 10.1109/TMC.2022.3226742. [DOI] [Google Scholar]
  • 62.Luo F., Khan S., Jiang B., Wu K. Vision Transformers for Human Activity Recognition using WiFi Channel State Information. IEEE Internet Things J. 2024;11:28111–28122. doi: 10.1109/JIOT.2024.3375337. [DOI] [Google Scholar]
  • 63.Zhang Y., Wang G., Liu H., Gong W., Gao F. WiFi-Based Indoor Human Activity Sensing: A Selective Sensing Strategy and a Multi-Level Feature Fusion Approach. IEEE Internet Things J. 2024;11:29335–29347. doi: 10.1109/JIOT.2024.3397708. [DOI] [Google Scholar]
  • 64.Niu K., Wang X., Zhang F., Zheng R., Yao Z., Zhang D. Rethinking Doppler Effect for Accurate Velocity Estimation With Commodity WiFi Devices. IEEE J. Sel. Areas Commun. 2022;40:2164–2178. doi: 10.1109/JSAC.2022.3155523. [DOI] [Google Scholar]
  • 65.Wu D., Zeng Y., Gao R., Li S., Li Y., Shah R.C., Lu H., Zhang D. WiTraj: Robust Indoor Motion Tracking With WiFi Signals. IEEE Trans. Mob. Comput. 2023;22:3062–3078. doi: 10.1109/TMC.2021.3133114. [DOI] [Google Scholar]
  • 66.Li W., Gao R., Xiong J., Zhou J., Wang L., Mao X., Yi E., Zhang D. WiFi-CSI Difference Paradigm: Achieving Efficient Doppler Speed Estimation for Passive Tracking. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;8:1–29. doi: 10.1145/3659608. [DOI] [Google Scholar]
  • 67.Zhang G., Zhang D., He Y., Chen J., Zhou F., Chen Y. Multi-Person Passive WiFi Indoor Localization With Intelligent Reflecting Surface. IEEE Trans. Wirel. Commun. 2023;22:6534–6546. doi: 10.1109/TWC.2023.3244369. [DOI] [Google Scholar]
  • 68.Zhang G., Zhang D., Deng H., Wu Y., Zhan F., Chen Y. Practical Passive Indoor Localization With Intelligent Reflecting Surface. IEEE Trans. Mob. Comput. 2024;23:12477–12490. doi: 10.1109/TMC.2024.3414861. [DOI] [Google Scholar]
  • 69.Fan Y., Zhang F., Wu C., Wang B., Liu K.J.R. RF-Based Indoor Moving Direction Estimation Using a Single Access Point. IEEE Internet Things J. 2022;9:462–473. doi: 10.1109/JIOT.2021.3083669. [DOI] [Google Scholar]
  • 70.Chi G., Yang Z., Xu J., Wu C., Zhang J., Liang J., Liu Y. Wi-drone: Wi-fi-based 6-DoF tracking for indoor drone flight control; Proceedings of the 20th Annual International Conference on Mobile Systems, Applications and Services (MobiSys ‘22); Portland, OR, USA. 27 June–1 July 2022; New York, NY, USA: Association for Computing Machinery; 2022. pp. 56–68. [Google Scholar]
  • 71.Zeng Y., Wu D., Xiong J., Liu J., Zhang D. MultiSense: Enabling Multi-person Respiration Sensing with Commodity WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020;4:1–29. doi: 10.1145/3411816. [DOI] [Google Scholar]
  • 72.Zhang F., Wu C., Wang B., Wu M., Bugos D., Zhang H., Liu K.J.R. SMARS: Sleep Monitoring via Ambient Radio Signals. IEEE Trans. Mob. Comput. 2021;20:217–231. doi: 10.1109/TMC.2019.2939791. [DOI] [Google Scholar]
  • 73.Yu B., Wang Y., Niu K., Zeng Y., Gu T., Wang L., Guan C., Zhang D. WiFi-Sleep: Sleep Stage Monitoring Using Commodity Wi-Fi Devices. IEEE Internet Things J. 2021;8:13900–13913. doi: 10.1109/JIOT.2021.3068798. [DOI] [Google Scholar]
  • 74.Liu J., Zeng Y., Gu T., Wang L., Zhang D. WiPhone: Smartphone-based Respiration Monitoring Using Ambient Reflected WiFi Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021;5:1–19. doi: 10.1145/3448092. [DOI] [Google Scholar]
  • 75.Hu J., Yang J., Ong J.-B., Wang D., Xie L. ResFi: WiFi-Enabled Device-Free Respiration Detection Based on Deep Learning; Proceedings of the 2022 IEEE 17th International Conference on Control & Automation (ICCA); Naples, Italy. 27–30 June 2022; pp. 510–515. [Google Scholar]
  • 76.Xie X., Zhang D., Li Y., Hu Y., Sun Q., Chen Y. Robust WiFi Respiration Sensing in the Presence of Interfering Individual. IEEE Trans. Mob. Comput. 2024;23:8447–8462. doi: 10.1109/TMC.2023.3348879. [DOI] [Google Scholar]
  • 77.Jiang W., Xue H., Miao C., Wang S., Lin S., Tian C., Murali S., Hu H., Sun Z., Su L. Towards 3D human pose construction using wifi; Proceedings of the 26th Annual International Conference on Mobile Computing and Networking (MobiCom ‘20); London, UK. 21–25 September 2020; New York, NY, USA: Association for Computing Machinery; 2020. pp. 1–14. [Google Scholar]
  • 78.Li C., Liu Z., Yao Y., Cao Z., Zhang M., Liu Y. Wi-fi see it all: Generative adversarial network-augmented versatile wi-fi imaging; Proceedings of the 18th Conference on Embedded Networked Sensor Systems (SenSys ‘20); Yokohama, Japan. 16–19 November 2020; New York, NY, USA: Association for Computing Machinery; 2020. pp. 436–448. [Google Scholar]
  • 79.Ren Y., Wang Z., Wang Y., Tan S., Chen Y., Yang J. GoPose: 3D Human Pose Estimation Using WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–25. doi: 10.1145/3534605. [DOI] [Google Scholar]
  • 80.Pallaprolu A., Korany B., Mostofi Y. Wiffract: A new foundation for RF imaging via edge tracing; Proceedings of the 28th Annual International Conference on Mobile Computing and Networking (MobiCom ‘22); Sydney, NSW, Australia. 17–21 October 2022; New York, NY, USA: Association for Computing Machinery; 2022. pp. 255–267. [Google Scholar]
  • 81.Zhou Y., Huang H., Yuan S., Zou H., Xie L., Yang J. MetaFi++: WiFi-Enabled Transformer-Based Human Pose Estimation for Metaverse Avatar Simulation. IEEE Internet Things J. 2023;10:14128–14136. doi: 10.1109/JIOT.2023.3262940. [DOI] [Google Scholar]
  • 82.Wang X., Niu K., Yu A., Xiong J., Yao Z., Wang J., Li W., Zhang D. WiMeasure: Millimeter-level Object Size Measurement with Commodity WiFi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2023;7:1–26. doi: 10.1145/3596250. [DOI] [Google Scholar]
  • 83.Yin C., Miao X., Chen J., Jiang H., Yang J., Zhou Y., Wu M., Chen Z. PowerSkel: A Device-Free Framework Using CSI Signal for Human Skeleton Estimation in Power Station. IEEE Internet Things J. 2024;11:20165–20177. doi: 10.1109/JIOT.2024.3369856. [DOI] [Google Scholar]
  • 84.Yao Z., Wang X., Niu K., Zheng R., Wang J., Zhang D. WiProfile: Unlocking Diffraction Effects for Sub-Centimeter Target Profiling Using Commodity WiFi Devices; Proceedings of the 30th Annual International Conference on Mobile Computing and Networking (ACM MobiCom ‘24); Washington, DC, USA. 18–22 November 2024; New York, NY, USA: Association for Computing Machinery; 2024. pp. 185–199. [Google Scholar]
  • 85.Wu D., Zeng Y., Zhang F., Zhang D. WiFi CSI-based device-free sensing: From Fresnel zone model to CSI-ratio model. CCF Trans. Pervasive Comput. Interact. 2022;4:88–102. doi: 10.1007/s42486-021-00077-z. [DOI] [Google Scholar]
  • 86.Niu K., Wang X., Yao Z., Zhang F., Cheng S., Jiang Y., Zhang D. How Target’s Location and Orientation Affect Velocity Extraction Accuracy in WiFi Sensing Systems; Proceedings of the ACM Turing Award Celebration Conference-China 2023 (ACM TURC ′23); Wuhan, China. 28–30 July 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 35–36. [Google Scholar]
  • 87.Zhang F., Jin B., Zhang D. Ubiquitous Wireless Sensing-Theory, Technique and Application; Proceedings of the ACM Turing Award Celebration Conference-China 2023 (ACM TURC ‘23); Wuhan, China. 28–30 July 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 33–34. [Google Scholar]
  • 88.Zhang J.A., Wu K., Huang X., Guo Y.J., Zhang D., Heath R.W. Integration of Radar Sensing into Communications with Asynchronous Transceivers. IEEE Commun. Mag. 2022;60:106–112. doi: 10.1109/MCOM.003.2200096. [DOI] [Google Scholar]
  • 89.Zeng Y., Wu D., Xiong J., Zhang D. Boosting WiFi Sensing Performance via CSI Ratio. IEEE Pervasive Comput. 2021;20:62–70. doi: 10.1109/MPRV.2020.3041024. [DOI] [Google Scholar]
  • 90.Zeng Y., Wu D., Xiong J., Yi E., Gao R., Zhang D. FarSense: Pushing the Range Limit of WiFi-based Respiration Sensing with CSI Ratio of Two Antennas. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019;3:1–26. doi: 10.1145/3351279. [DOI] [Google Scholar]
  • 91.Zeng Y., Liu J., Xiong J., Liu Z., Wu D., Zhang D. Exploring Multiple Antennas for Long-range WiFi Sensing. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;5:1–30. doi: 10.1145/3494979. [DOI] [Google Scholar]
  • 92.Li Y., Wu D., Zhang J., Xu X., Xie Y., Gu T., Zhang D. DiverSense: Maximizing Wi-Fi Sensing Range Leveraging Signal Diversity. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–28. doi: 10.1145/3536393. [DOI] [Google Scholar]
  • 93.Wang X., Niu K., Xiong J., Qian B., Yao Z., Lou T., Zhang D. Placement Matters: Understanding the Effects of Device Placement for WiFi Sensing. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022;6:1–25. doi: 10.1145/3517237. [DOI] [Google Scholar]
  • 94.Xie Y., Li Z., Li M. Precise Power Delay Profiling with Commodity WiFi; Proceedings of the 21st Annual International Conference on Mobile Computing and Networking (MobiCom ‘15); Paris, France. 7–11 September 2015; New York, NY, USA: Association for Computing Machinery; 2015. pp. 53–64. [Google Scholar]
  • 95.Gringoli F., Schulz M., Link J., Hollick M. Free Your CSI: A Channel State Information Extraction Platform For Modern Wi-Fi Chipsets; Proceedings of the 13th International Workshop on Wireless Network Testbeds, Experimental Evaluation & Characterization (WiNTECH ‘19); Los Cabos, Mexico. 25 October 2019; New York, NY, USA: Association for Computing Machinery; 2019. pp. 21–28. [Google Scholar]
  • 96.Hernandez S.M., Bulut E. Lightweight and Standalone IoT Based WiFi Sensing for Active Repositioning and Mobility; Proceedings of the 2020 IEEE 21st International Symposium on “A World of Wireless, Mobile and Multimedia Networks” (WoWMoM); Cork, Ireland. 31 August–3 September 2020; pp. 277–286. [Google Scholar]
  • 97.Hernandez S.M., Bulut E. WiFi Sensing on the Edge: Signal Processing Techniques and Challenges for Real-World Systems. IEEE Commun. Surv. Tutorials. 2023;25:46–76. doi: 10.1109/COMST.2022.3209144. [DOI] [Google Scholar]
  • 98.Gringoli F., Cominelli M., Blanco A., Widmer J. AX-CSI: Enabling CSI Extraction on Commercial 802.11ax Wi-Fi Platforms; Proceedings of the 15th ACM Workshop on Wireless Network Testbeds, Experimental evaluation & CHaracterization (WiNTECH ‘21); New Orleans, LA, USA. 31 January–4 February 2022; New York, NY, USA: Association for Computing Machinery; 2021. pp. 46–53. [Google Scholar]
  • 99.Jiang Z., Luan T.H., Ren X., Lv D., Hao H., Wang J., Zhao K., Xi W., Xu Y., Li R. Eliminating the Barriers: Demystifying Wi-Fi Baseband Design and Introducing the PicoScenes Wi-Fi Sensing Platform. IEEE Internet Things J. 2022;9:4476–4496. doi: 10.1109/JIOT.2021.3104666. [DOI] [Google Scholar]
  • 100.Yousefi S., Narui H., Dayal S., Ermon S., Valaee S. A Survey on Behavior Recognition Using WiFi Channel State Information. IEEE Commun. Mag. 2017;55:98–104. doi: 10.1109/MCOM.2017.1700082. [DOI] [Google Scholar]
  • 101.Ma Y., Zhou G., Wang S., Zhao H., Jung W. SignFi: Sign Language Recognition Using WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;2:1–21. doi: 10.1145/3191755. [DOI] [Google Scholar]
  • 102.Palipana S., Rojas D., Agrawal P., Pesch D. FallDeFi: Ubiquitous Fall Detection using Commodity Wi-Fi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018;1:1–25. doi: 10.1145/3161183. [DOI] [Google Scholar]
  • 103.Guo L., Guo S., Wang L., Lin C., Liu J., Lu B., Fang J., Liu Z., Shan Z., Yang J. Wiar: A Public Dataset for Wifi-Based Activity Recognition. IEEE Access. 2019;7:154935–154945. doi: 10.1109/ACCESS.2019.2947024. [DOI] [Google Scholar]
  • 104.Zheng Y., Zhang Y., Qian K., Zhang G., Liu Y., Wu C., Yang Z. Zero-Effort Cross-Domain Gesture Recognition with Wi-Fi; Proceedings of the 17th Annual International Conference on Mobile Systems, Applications and Services (MobiSys ′19); Seoul, Republic of Korea. 17–21 June 2019; New York, NY, USA: Association for Computing Machinery; 2019. pp. 313–325. [Google Scholar]
  • 105.Xiao R., Liu J., Han J., Ren K. OneFi: One-Shot Recognition for Unseen Gesture via COTS WiFi; Proceedings of the 19th ACM Conference on Embedded Networked Sensor Systems (SenSys ‘21); Coimbra, Portugal. 15–17 November 2021; New York, NY, USA: Association for Computing Machinery; 2021. pp. 206–219. [Google Scholar]
  • 106.Yang J., Chen X., Zou H., Lu X., Wang D., Yang S.J., Huang H., Zhou Y., Chen X., Xu Y., et al. MM-Fi: Multi-Modal Non-Intrusive 4D Human Dataset for Versatile Wireless Sensing. In: Oh A., Naumann T., Globerson A., Saenko K., Hardt M., Levine S., editors. Advances in Neural Information Processing Systems. Volume 36. Curran Associates, Inc.; Red Hook, NY, USA: 2023. pp. 18756–18768. [Google Scholar]
  • 107.Xie S., Xie L. SenseFi: A library and benchmark on deep-learning-empowered WiFi human sensing. Patterns. 2023;4:100703. doi: 10.1016/j.patter.2023.100703. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Cominelli M., Gringoli F., Restuccia F. Exposing the CSI: A Systematic Investigation of CSI-based Wi-Fi Sensing Capabilities and Limitations; Proceedings of the 2023 IEEE International Conference on Pervasive Computing and Communications (PerCom); Atlanta, GA, USA. 13–17 March 2023; pp. 81–90. [Google Scholar]
  • 109.Yi E., Zhang F., Xiong J., Niu K., Yao Z., Zhang D. Enabling WiFi Sensing on New-generation WiFi Cards. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;7:1–26. doi: 10.1145/3633807. [DOI] [Google Scholar]
  • 110.Yang Z., Zhang Y., Qian K., Wu C. SLNet: A Spectrogram Learning Neural Network for Deep Wireless Sensing; Proceedings of the 20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23); Boston, MA, USA. 17–19 April 2023; Berkeley, CA, USA: USENIX Association; 2023. pp. 1221–1236. [Google Scholar]
  • 111.Chi G., Yang Z., Wu C., Xu J., Gao Y., Liu Y., Han T.X. RF-Diffusion: Radio Signal Generation via Time-Frequency Diffusion; Proceedings of the 30th Annual International Conference on Mobile Computing and Networking (ACM MobiCom ‘24); Washington, DC, USA. 18–22 November 2024; New York, NY, USA: Association for Computing Machinery; 2024. pp. 77–92. [Google Scholar]
  • 112.Hou W., Wu C. RFBoost: Understanding and Boosting Deep WiFi Sensing via Physical Data Augmentation. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024;8:1–26. doi: 10.1145/3659620. [DOI] [Google Scholar]
  • 113.He Y., Liu J., Li M., Yu G., Han J. Forward-Compatible Integrated Sensing and Communication for WiFi. IEEE J. Sel. Areas Commun. 2024;42:2440–2456. doi: 10.1109/JSAC.2024.3413955. [DOI] [Google Scholar]
  • 114.Wu C., Huang X., Huang J., Xing G. Enabling Ubiquitous WiFi Sensing with Beamforming Reports; Proceedings of the ACM SIGCOMM 2023 Conference (ACM SIGCOMM ‘23); New York, NY, USA. 10 September 2023; New York, NY, USA: Association for Computing Machinery; 2023. pp. 20–32. [Google Scholar]
  • 115.Yi E., Wu D., Xiong J., Zhang F., Niu K., Li W., Zhang D. BFMSense: WiFi Sensing Using Beamforming Feedback Matrix; Proceedings of the 21st USENIX Symposium on Networked Systems Design and Implementation (NSDI24); Santa Clara, CA, USA. 16–18 April 2024; Berkeley, CA, USA: USENIX Association; 2024. pp. 1697–1712. [Google Scholar]
  • 116.Korany B., Karanam C.R., Cai H., Mostofi Y. XModal-ID: Using WiFi for Through-Wall Person Identification from Candidate Video Footage; Proceedings of the 25th Annual International Conference on Mobile Computing and Networking (MobiCom ‘19); Los Cabos, Mexico. 21–25 October 2019; New York, NY, USA: Association for Computing Machinery; 2019. pp. 1–15. [Google Scholar]
  • 117.De Sanctis M., Domenico S.D., Fioravanti D., Abellan E.B., Rossi T., Cianca E. RF-Based Device-Free Counting of People Waiting in Line: A Modular Approach. IEEE Trans. Veh. Technol. 2022;71:10471–10484. doi: 10.1109/TVT.2022.3182548. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES