| A3C | Asynchronous Advantage Actor Critic |
| ACER | Actor Critic with Experience Replay |
| ACK | Acknowledgement |
| AH | Always High |
| AL | Always Low |
| AS | Attacking Sensor |
| CDF | Cumulative Distribution Function |
| CM | Cramer-von Mises |
| CNNME | Communication and Neural Networks Mean Embedding |
| CMME | Communication and Mean-based Mean Embedding |
| DLA | Deep Learning Attacker |
| DNN | Deep Neural Network |
| DoS | Denial of Service |
| DRL | Deep Reinforcement Learning |
| DRQN | Deep Recurrent Q-Networks |
| DQN | Deep Q-Networks |
| FC | Fusion Center |
| GS | Good Sensor |
| MAC | Medium Access Control |
| MDP | Markov Decision Process |
| ME | Mean Embedding |
| MME | Mean-based Mean Embedding |
| NNME | Neural Networks Mean Embedding |
| PHY | Physical |
| POMDP | Partially Observable Markov Decision Process |
| PPO | Proximal Policy Optimization |
| QoS | Quality of Service |
| RL | Reinforcement Learning |
| RNN | Recurrent Neural Network |
| RP | Random Policy |
| SNR | Signal-to-Noise Ratio |
| SSDF | Spectrum Sensing Data Falsification |
| TRPO | Trust Region Policy Optimization |
| WSN | Wireless Sensor Network |