Abstract
The motivation for this article stems from the fact that medical image security is crucial for maintaining patient confidentiality and protecting against unauthorized access or manipulation. This paper presents a novel encryption technique that integrates the Deep Convolutional Generative Adversarial Networks (DCGAN) and Virtual Planet Domain (VPD) approach to enhance the protection of medical images. The method uses a Deep Learning (DL) framework to generate a decoy image, which forms the basis for generating encryption keys using a timestamp, nonce, and 1-D Exponential Chebyshev map (1-DEC). Experimental results validate the efficacy of the approach in safeguarding medical images from various security threats, including unauthorized access, tampering, and adversarial attacks. The randomness of the keys and encrypted images are demonstrated through the National Institute of Standards and Technology (NIST) SP 800-22 Statistical test suite provided in Tables 4 and 14, respectively. The robustness against key sensitivity, noise, cropping attacks, and adversarial attacks are shown in Figs. 15–18, 22–23, and 24. The data presented in Tables 5, 6, and 7 shows the proposed algorithm is robust and efficient in terms of time and key space complexity. Security analysis results are shown (such as histogram plots in Figs. 11–14 and correlation plots in Figs. 19–21). Information Entropy (
), correlation coefficient (
), Mean Square Error (MSE) (
), Peak Signal to Noise Ratio (PSNR) (
), Number of Pixel Change Rate (NPCR) (
), and Unified Average Changing Intensity (UACI) (
) underscore the high security and reliability of the encrypted images, are shown in Tables 8–11. Further, statistical NPCR and UACI are calculated in Tables 12 and 13, respectively. The proposed algorithm is also compared with existing algorithms, and compared values are provided in Table 15. The data presented in Tables 3–15 suggest that the proposed algorithm can opt for practical use.
Keywords: DCGAN, Chaotic map, VPD, Encryption, Decryption
Subject terms: Computer science, Applied mathematics, Information technology, Machine learning, Image processing
Introduction
Medical image encryption is vital for maintaining the confidentiality and integrity of sensitive medical data, ensuring that it remains accessible only to authorized individuals. Protecting the privacy and integrity of medical images is becoming increasingly important as medical imaging technology advances, improving diagnostic accuracy and guiding treatment plans. Medical images, such as X-rays, Magnetic Resonance Images (MRI), and Computed Tomography (CT) scans, contain highly personal information that could be misused if accessed by unauthorized parties, leading to privacy violations and even potential financial or reputational harm to patients. Furthermore, the precision of medical images is essential for correct diagnosis and treatment, as unauthorized alteration or corruption of these images could lead to incorrect treatment plans, potentially endangering patient health.
The growing adoption of telemedicine and cloud-based healthcare systems has further emphasized the need for secure data transmission and storage. Encryption ensures that medical images transmitted over networks or stored in cloud platforms remain protected from breaches and cyber threats while facilitating research and collaboration by safeguarding patient privacy.
In the literature, various sophisticated encryption algorithms have been developed in1–24 and references therein to secure medical images, each with strengths and limitations. DNA encoding offers robust encryption by leveraging genetic sequence patterns but is constrained by the need for specialized setups, limiting its practicality1–4. Chaotic maps, often combined with substitution boxes, are valued for their sensitivity to initial conditions and nonlinearity, which are core features of chaos theory, making it ideal for cryptographic applications5–7. Digital chaos, modeled as iterative dynamical systems, can be implemented across one-dimensional (1D), two-dimensional (2D), and hyperdimensional (HD) frameworks. These have been applied to domains such as secure communication25, pseudo-random number generation26, image encryption27,28, and video encryption29,30. While higher-dimensional chaotic systems offer larger key spaces and more complex behaviors for enhanced security, they also incur higher computational costs and increased encryption time. Conversely, 1D chaotic systems are computationally efficient but may exhibit limitations like dynamic degeneracy and non-uniform distributions, potentially reducing their unpredictability. Techniques such as zigzag XORing8–11 and block-based methods like AES12,13 provide reliable security, but AES struggles with the strong pixel correlations in medical images, leading to inefficiencies. Deep learning (DL) has revolutionized encryption, enabling models like Convolutional Neural Networks (CNNs) to dynamically learn encryption patterns, manage keys, and detect anomalies, enhancing adaptability and resilience against attacks31–35. Methods such as DeepEDN36, EncipherGAN15, and ResNet-based approaches 16 integrate DL with traditional cryptographic techniques for enhanced robustness. Dynamic key generation methods that leverage neural networks have also been proposed, which adapt to the needs of the dynamic and scalable encryption necessary for modern medical imaging systems in19. Novel DL applications, including direct image encryption using DCT coefficients37, DCGANs with quaternion mathematics17, and GANs generating synthetic images for privacy18, highlight the field’s innovation. However, these methods often face challenges like high computational demands and scalability issues, particularly during encryption and decryption. Hybrid approaches combining chaotic maps, GANs, and Huffman compression20, as well as 4D chaotic systems for critical area protection21, offer improved multilayered security but remain computationally expensive. Addressing these limitations requires balancing computational efficiency, adaptability, and security, paving the way for more practical and scalable encryption solutions in medical imaging applications.
Motivated by the limitations and challenges described above, we propose a novel encryption method for medical images that employs a DCGAN associated with VPD to increase the efficiency of cryptographic operations by segregating the key generation phase from the encryption and decryption routines, thus reducing computational load. Furthermore, chaotic maps are utilized to derive robust encryption keys, with security enhancements achieved through techniques such as VPD, 3D pixel intershuffling, and zigzag XORing. The proposed methodology seeks to balance enhanced security with computational efficiency, addressing contemporary cryptographic application evolving needs and significantly complicating unauthorized data access or manipulation.
Organization of the work
"Preliminaries" provides a detailed discussion of the DCGAN, 1-DEC, and VPD. In "Key generation procedure", the key generation procedure is explained along with pseudocode. "Proposed encryption and decryption process" presents the proposed encryption and decryption algorithms and pseudocodes, including encryption and decryption results. In "Security analysis", security analysis reports of the proposed algorithm are given. We provide a comparative analysis with existing methods in " Comparative analysis ". The last section concludes the proposed work.
Preliminaries
In this section, we discuss three main ingredients (DCGAN, 1-DEC, and VPD) used to generate random keys and employ them during the encryption process to provide robust support for the proposed novel encryption algorithm.
DCGAN
The GAN38 is a deep learning architecture where two neural networks, the generator
and the discriminator
, are involved in a game theoretic interaction to produce realistic synthetic data. The generator receives a noise vector
as input, typically sampled from a prior distribution such as either the uniform or the Gaussian noise distribution. The generator’s objective is to transform
into synthetic data that closely matches the training data. The discriminator takes samples drawn from the given training data
and from the generated synthetic data
as input and attempts to classify them as real (label 1) or fake (label 0). In a min-max game between the two models with opposing objectives - the discriminator aims to get better at distinguishing between real and fake images, whereas the generator learns to produce fake images that look like the original data, ultimately ‘fooling’ the discriminator into classifying a fake image as real.
As a novel image generation model, GANs have gained popularity among researchers, leading to the development of numerous generative deep learning frameworks. Key breakthroughs include CGAN39for stable training with category labels, DCGAN with improved training approaches40, and subsequent advancements such as Pix2Pix41, CycleGAN42, StyleGAN43, and BigGAN44, which have optimized the image generation in a variety of real-world settings. DCGANs45 integrated CNNs into the GAN architecture. Applications in image generation46 reveal that DCGAN has produced images that are up to twice as high in quality compared to those generated by standard GANs, along with significantly greater diversity in images, establishing themselves as powerful techniques for generative learning. Its architecture is versatile enough to generate specific types of decoy images, enabling the model to be tailored to meet the unique needs of an encryption scheme.
In47, various deep learning techniques used in image encryption are surveyed. The use of deep learning enhances the encryption system’s security, particularly against plaintext attacks, through non-linearity in cryptanalysis. One common approach is style transfer, where GANs and CycleGANs are used to convert plaintext images into ciphertexts. The encryption network is trained to generate ciphertext images, while the decryption network aims to restore the original image. Here, the parameters of the networks act as the secret keys, resulting in a large key space that increases security. In a medical imaging application for diagnosis and treatments based on brain MRI and CT scan data36, introduces “DLEDNet,” a deep learning architecture where encryption and decryption networks, along with a discriminator neural network, combine to enhance the encryption performance. Similarly, in15, a cycle-GAN-based algorithm is developed for image encryption, and tested on medical datasets. Compared to chaotic-map-based encryption methods, which are vulnerable to phase-space-reconstruction attacks, the proposed model’s nonlinearity enhances the encryption system’s security. Style transfer approaches with additional diffusion properties, as seen in48, improve the robustness of encryption systems. Here, a cover image is used to disguise the plaintext, followed by a domain transformation using CycleGAN, combining steganography with deep learning. Chaotic sequences also play a significant role in deep learning-based encryption, where convolutional kernels are generated by chaotic sequences. As noted in37, chaotic maps used in the diffusion of the plaintext image result in dynamic key generation, which can be combined with traditional encryption techniques like XOR49.
The main improvements in DCGAN over the classic GAN include replacing pooling functions with strided and fractional-strided convolutions50, implementing batch normalization51 across all layers, removing fully connected layers to simplify the architecture, and using the Rectified Linear Unit (ReLU)52 and tanh activations in the generator and leaky ReLU in the discriminator, all of which contribute to improved stability and image quality. The generator takes input from a latent space (z) and passes it through a series of fractional-strided convolutional-transpose layers to produce synthetic images. The discriminator is a binary CNN classifier that uses strided convolutions to downsample the input images into categories (real/fake). The detailed architecture and training process are discussed in the following sections.
Strided convolutions and fractional-strided convolutions
Convolutional layers apply a kernel
, which is like a filter that slides over an input layer
to extract spatial features,
, called feature maps. The mathematical operation is defined as follows:
![]() |
where
are spatial indices, and
denote the kernel dimensions. Strided convolutions introduce a step variable (or stride) that controls the number of pixels the kernel skips while moving along the width and height of the input; downsampling the feature map can be seen in Fig. 3. This reduction in spatial dimensions helps the discriminator focus on global features, such as object shapes. A padding of zeros is added to preserve information on the border pixels. The size of the output image obtained after convolution is given by:
![]() |
Fractionally-strided convolutions (also known as transposed convolutions) reverse the downsampling process to upsample input data. By inserting zeros between entries of the input feature maps, they increase the spatial dimensions of the feature map, effectively creating higher-resolution images. This upsampling is crucial for the generator to create large, high-quality images from a small latent vector. The output size is formulated as follows:
![]() |
This process can be visualized in Fig. 3.
Fig. 3.

Elements of the convolutional layers: (a) Strided convolution operation (b) Fractional-strided convolution operation (c) Batch normalization layer.
Batch normalization
Batch normalization is a neural network layer, added after a hidden layer, that normalizes the output of each layer
using batch mean
and variance
, along with learnable parameters
and
, before passing the result to the next layer (Fig. 3). This technique stabilizes training by reducing internal covariate shifts and helps the model converge faster by ensuring the inputs to each layer maintain a consistent distribution. The absence of batch normalization can cause shifts in the distribution of layer inputs during training, potentially slowing down learning and causing instability. The normalized output is given by:
![]() |
Activation functions
Activation functions add non-linearity to neural networks, enabling them to model complex relationships between inputs and outputs. The generator uses ReLU in its hidden layers to prevent gradient saturation, while Leaky ReLU in the discriminator ensures better gradient flow for negative inputs, aiding effective learning. The tanh function scales the generator’s output to
, aligning with normalized training data, and the sigmoid function
in the discriminator’s final layer maps outputs to probabilities
, facilitating classification of real versus fake inputs. These activation functions, summarized in Table 1, play a critical role in the DCGAN architecture.
Table 1.
Activation functions used in DCGAN neural network layers.
| Activation function | Formula |
|---|---|
| Sigmoid | ![]() |
| Hyperbolic Tangent (Tanh) | ![]() |
| ReLU (Rectified Linear Unit) | ![]() |
| Leaky ReLU | ![]() |
Training process
The generator and discriminator are trained adversarially using the following loss functions:
- Generator Loss: Encourages
to generate realistic images: 
- Discriminator Loss: Encourages
to distinguish between real and fake images: 
These functions can be combined into a unified optimization function
:
![]() |
The DCGAN training alternates between optimizing
for several steps and then updating
for one step. In each iteration,
is trained to improve its accuracy by adjusting its weights via backpropagation. After
is trained and optimized for a given
,
receives the feedback from
’s loss function and updates its parameters to produce data that better approximates the real data distribution. This iterative approach continues with
being optimized again while
is fixed. As the training progresses,
indirectly learns the data distribution, gradually degrading
’s performance. For a fixed
, the optimal
converges to:
![]() |
where
is the training data distribution and
is the generative distribution of samples
. As training advances with sufficient capacity and iterations, the system converges to
approaching 0.5 for all
, meaning the discriminator can no longer differentiate between the two classes. This is the Nash equilibrium53 that occurs when
, indicating that the generator is successful in learning the probability distribution of the training dataset and can now be used to generate new images. In the proposed key generation, the trained generator in a DCGAN is used to generate the synthetic images that serve as the basis for creating encryption keys for input medical data. It is trained in advance, ensuring the generator is readily available for immediate key generation, thereby eliminating any delays in the key generation workflow. The unpredictable nature of synthetic images produced from noise and the complex non-linear properties of the deep CNN provides an additional level of security to the key space generated and the proposed encryption algorithm.
Experiments
This section provides detailed descriptions of the DCGAN’s layers and parameters. Details about the DCGAN’s design, including the number of convolutional layers used, activation functions, and other important hyperparameters, like the size of the latent vector, batch size, learning rate, and optimizer settings, along with their impact on model performance, are given. This information is necessary for experimental evaluation.
DCGAN architecture and model parameters
The DCGAN used in our paper is designed to produce RGB images of
dimensions, consistent with the image size of the dataset utilized in this work (described in a later section). The generator begins with an input noise vector
of size
sampled from a Gaussian distribution (Fig. 4). The generator’s core is organized into four groups of layers, each comprising a transposed convolution followed by batch normalization and a ReLU activation function. The final layer of the generator applies a tanh activation function to produce the output image. The input noise vector
is initially reshaped into a
grid with 1024 feature maps. This vector is upsampled through the convolutional layers, gradually producing a high-resolution image with dimensions
. The deconvolution step is set to 2. Therefore, each output of a group augments fourfold the input, resulting in the output sizes of layers being
,
,
,
, and finally
in the final layer. The number of feature maps is also set to 512, 256, 128, and 64, respectively, with the final layer having 3 channels. The discriminator mirrors the generator but with strided convolutions instead of transposed convolutions. The input, either a real image or a synthetic image from the generator, is progressively downsampled through these convolutions that decrease the spatial dimensions while increasing the depth of the feature maps. The discriminator consists of three groups of layers, each containing a strided convolution followed by batch normalization (applied to all layers except the input) and a Leaky ReLU activation function. The moving step of the convolution kernel is set to 2. For an input image of dimensions
, the output sizes of the convolution layers are
,
,
, and
, with 64, 128, 256, and 512 feature maps, respectively. The final layer is a fully connected layer used for flattening feature maps into a single output neuron with a sigmoidal activation function, thus producing a probability score between 0 and 1. The score refers to how confident the model is in its classification of the input image as real or fake. Figure 1 represents the complete DCGAN architecture and training flow, Fig. 2 illustrates the learned representations of inputs in the generator and discriminator, and the layers of the neural network are shown in Fig. 3.
Fig. 4.

Distribution of noise vector
following a Gaussian function.
is the input to the generator network of the DCGAN.
Fig. 1.
Architecture of the DCGAN consisting of generator network, discriminator network, and game theoretical min-max loss function.
Fig. 2.
Learned representations of inputs in (a) the generator and (b) the discriminator of the DCGAN.
Model hyperparameter settings
In our experiments, we generally followed the default hyperparameter settings proposed in the original DCGAN publication45. The learning rate
was set to 0.0002 and the batch to 64. The Adam optimizer was selected for its adaptive learning rate.
, known as the First Moment Decay Rate, was set to 0.5; this helps prevent oscillations and inconsistent updates during training. The
is the Second Moment Decay Rate, which controls the decay rate of the moving average for the squares of the gradients. It was kept at 0.999, which prevents the model from being too reactive to small changes in gradients and helps in more stable convergence. For the discriminator, the binary cross-entropy function was used to calculate loss. The generator network parameters are updated after one step of discriminator optimization. We have taken 100 epochs for simulation purposes. Users can select it as they see fit and use it as a secret key. The final image outputs from the generator were then used to create the initial keys for encryption and decryption in the proposed security framework.
The generator and discriminator training loss curves across 30,000 iterations are shown in Fig. 5. During the first 3000 iterations, the discriminator’s loss rapidly decreases while the generator’s loss is initially high. Both losses stabilize after approximately 3000 iterations, indicating that the generator begins creating more realistic images, making it difficult for the discriminator to differentiate between real and fake data. After this point, the losses fluctuate but remain relatively stable, suggesting a balanced training process. This indicates the model is generating useful outputs, and further hyperparameter optimization could enhance this stability. Training could be stopped when significant divergence is observed again, around 5000 iterations, as further training might lead to instability. Figure 6 shows the comparison between training images and generated images.
Fig. 5.
Training losses of Generator
and Discriminator
showing the convergence conditions of DCGAN.
Fig. 6.
(a) Skin cancer images in the training dataset, (b) Images produced by the Generator of the DCGAN.
System specifications for model training
The model training was conducted on an HPC server with an NVIDIA RTX A6000 GPU, offering 49140 MiB of dedicated memory for high-performance computational tasks. The server’s CPU is an AMD Ryzen 9 7950X 16-core Processor, supported by 125 GB of RAM and an additional 279 GB of swap memory available to handle overflow. The storage capacity includes three high-speed NVMe drives: nvme0n1p3 with 458 GB (27 GB used) and two drives (nvme2n1p1 and nvme1n1p1), each offering 3.6 TB of storage. The training environment utilized PyTorch version 2.4.0+cu121, which supports CUDA 12.1. This setup provided the computational power necessary to handle the extensive operations involved in deep learning tasks, particularly the iterative training process of GANs.
1-DEC
The 1-DEC was proposed in54 to overcome the traditional 1-Dimensional Chebyshev Chaotic Map (1-DCCM), tent map, and logistic map that encounter limitations related to its control parameters, empty intervals, and non-uniform state distribution. Adding an exponential function to the 1-DCCM greatly improved its performance, and it is called 1-DEC. The 1-DEC map is expressed in Eq. (1) as follows:
![]() |
1 |
where
is the control parameter of the 1-DEC, and
represents the output sequence generated by the map. The sequence values are uniformly distributed, as discussed in54.
It must be pointed out that degeneration is a significant challenge in discrete chaotic maps used for cryptographic applications. It occurs when chaotic systems lose their inherent randomness due to limitations of finite precision in digital implementations, leading to periodic or fixed-point dynamics and degradation in key properties such as sensitivity to initial conditions. Solutions to this problem, such as improving computation precision, adopting hyperchaotic maps, or employing multiple chaotic systems, have been proposed. However, increasing precision or combining multiple systems can incur higher computational costs and complexity in implementation, which may hinder their practicality. The 1-DEC map combines trigonometric and exponential functions to achieve highly nonlinear dynamics, making it less degenerate than traditional 1D maps. These dynamics ensure that even under finite precision, the map maintains its chaotic properties, including sensitivity to initial conditions and ergodicity. This is demonstrated in a comparative analysis with traditional 1D maps like the Chebyshev map, the logistic map, and other recent maps 1-DCP55, and 1-DSP56.
Bifurcation diagram
Considering the initial condition
(i.e.,
), iterating the 1-DEC for 150 cycles while altering the control parameter
yields a sequence of length 150. To mitigate transient effects, the last 100 output values are chosen for 2D plot generation. The findings suggest that, regardless of fluctuations in the control parameter
, the sequence values generated by the 1-DEC show a uniform distribution across the entire space, spanning the complete range of mapping states. The same is true for the initial conditions
and
. Bifurcation diagrams are essential in both theoretical and applied dynamics, offering a clear visual view to explore and understand complex behaviors in nonlinear systems as parameters change. The bifurcation diagrams in Fig. 7 depict the behavior of the 1-DEC map under different initial conditions (
) with the control parameter
shown over the interval
. These diagrams demonstrate that the sequence values produced by 1-DEC remain uniformly distributed across the range of mapping states, irrespective of fluctuations in the initial conditions or control parameter values within this range. This uniformity confirms the map’s strong chaotic properties and supports its utility in encryption, as it enhances resistance to predictability and plain-image attacks. In contrast, Fig. 8 highlights the limitations of simpler maps like the logistic and Chebyshev maps. The logistic map demonstrates chaotic behavior only within a narrow range of
(close to
to
) but converges to fixed points or periodic orbits outside this interval, reducing randomness. The Chebyshev map fails to exhibit chaos for
, limiting its applicability. In a similar fashion, the 1-DCP and 1-DCP maps are also not consistently chaotic for all intervals of control parameters. The comparisons underscore the superiority of the 1-DEC map, as its bifurcation diagrams consistently demonstrate uniformity and robust chaotic behavior across a broader range of parameters and initial conditions.
Fig. 7.
Bifurcation diagrams of the 1-DEC with varying initial conditions: (a) 0.25454, (b) 0.1, (c) 0.4.
Fig. 8.
Bifurcation diagrams of (a) the Chebyshev map, (b) the logistic map, (c) 1-DCP map, and (d) 1-DSP map (control parameter
) with initial condition
.
Lyapunov Exponent
The Lyapunov Exponent (LE) quantifies the average rate of divergence or convergence of two infinitesimally close trajectories in the state space of a chaotic map, determining its sensitivity to initial conditions. It is mathematically expressed as:
![]() |
where
is the derivative of the map at the
-th iteration. A positive LE (
) across a wide range of control parameters indicates strong chaotic behavior and resilience against degeneration. The 1-DEC map’s superior LE (
across its parameter range) confirms its robust chaotic behavior, as trajectories diverge exponentially even in discrete implementations as seen in comparative analyses with the other maps in Fig. 9.
Fig. 9.

Lyapunov exponent analysis and comparison.
The key advantage of employing the 1-DEC map lies in its straightforward hardware implementation. While higher-dimensional maps57–60 can offer increased complexity, they demand a greater number of logic gates, are often computationally expensive and pose challenges in precision management, as errors propagate more rapidly in the multi-dimensional system. One-dimensional maps provide a more efficient and less complex alternative. The proposed encryption algorithm uses a 1-DEC map that produces satisfactory results and is immune to plain-image attacks.
VPD
To address the limitations of traditional DNA coding, a novel solution is the VPD encoding scheme published in61. This innovative approach creates 40320-rules for XORing with eight virtual planets: Uranus (Ur), Mars (Ma), Venus (Ve), Neptune (Ne), Jupiter (Ju), Earth (Ea), Saturn (Sa), and Mercury (Me). Each of these planets is uniquely identified by a 3-bit binary number in Table 2, allowing eight distinctive combinations. The VPD process involves applying the XOR operation, precisely a bitwise-XOR operation, among different planets. For instance, considering Venus (Ve) and Earth (Ea), which are represented as 010 and 101 in binary, the XOR operation yields 111, corresponding to the planet Mercury (Me). This standardized approach introduces a creative and systematic method for encoding binary information via a virtual solar system. The proposed approach uses Rule-37241 from61, which is listed in Table 3 as follows:
Table 2.
Planets bitwise representation.
| Planets | Ur | Ma | Ve | Ne | Ju | Ea | Sa | Me |
|---|---|---|---|---|---|---|---|---|
| Bit representation | 000 | 001 | 010 | 011 | 100 | 101 | 110 | 111 |
Table 3.
A comprehensive table illustrating the XOR operation between each virtual planet, utilizing Rule-37241, is shown in Table 2.
![]() |
Ur | Ma | Ve | Ne | Ju | Ea | Sa | Me |
|---|---|---|---|---|---|---|---|---|
| Ur | Ur | Ma | Ve | Ne | Ju | Ea | Sa | Me |
| Ma | Ma | Ur | Ne | Ve | Ea | Ju | Me | Sa |
| Ve | Ve | Ne | Ur | Ma | Sa | Me | Ju | Ea |
| Ne | Ne | Ve | Ma | Ur | Me | Sa | Ea | Ju |
| Ju | Ju | Ea | Sa | Me | Ur | Ma | Ve | Ne |
| Ea | Ea | Ju | Me | Sa | Ma | Ur | Ne | Ve |
| Sa | Sa | Me | Ju | Ea | Ve | Ne | Ur | Ma |
| Me | Me | Sa | Ea | Ju | Ne | Ve | Ma | Ur |
The motivation for utilizing the 1-DEC and VPD domains in the proposed encryption algorithm is to achieve heightened key sensitivity and increased randomness, thereby bolstering the algorithm’s resilience against cryptographic attacks, particularly side-channel attacks, and ease of implementation on hardware (IoT devices).
Key generation procedure
This section discusses the key generation procedure in detail, presents the randomness test results, and provides the corresponding pseudocode. The keys generated by the proposed method are verified and added through the NIST SP 800-22 statistical test suite to ensure their randomness. We use DCGAN to generate a random, fake-color image for key generation. The generated image is similar to the original image but can have very different RGB values. The base data set used for DCGAN training, which includes more than 10,000 images of skin cancer, is sourced from Kaggle (https://www.kaggle.com/code/jnegrini/ham10000-analysis-and-model-comparison). The architecture of the DCGAN involves a generator and discriminator, as explained in “DCGAN”. The training process requires multiple iterations to optimize the generator and discriminator networks until the generator can produce realistic synthetic images that closely mimic the distribution of the training data. The trained generator is ready for real-time key generation without additional computational overhead. We also used the time-stamp format: DD-MM-YYYY HH:MM:SS (Day-Month-Year Hour:Minute:Second) along with nonce during the key generation process along with 1-DEC; the steps are as follows:
Step 1: Read the DCGAN-generated image; separate it into its red, green, and blue channels; and calculate the sum of pixels for each channel. After calculating the sum of the pixel values for each color channel, the code constructs a string sum containing the sums of the red, green, and blue channels incremented by 1, converted to strings. Additionally, the sum string is concatenated with a timestamp and a nonce. This step essentially supports the key space if all the pixel values are zero.
Step 2: Sum converted to 512 Hash through SHA-512.
Step 3: We divide the number into eight parts, convert every part into a decimal of the 15 digits, and divide it by
.Step 4: The initial condition for the 1-DEC is obtained in Step 3, which iteratively generates eight sequences of data streams. These streams are used in the proposed encryption algorithm to perform shuffle and plane XOR. We consider them keys (
).
The key generation procedure is further explained through pseudocode, which is provided in Algorithm 1.
Algorithm 1.
Key generation procedure
Randomness results on keys (
)
We generated 12,000,000 bits and divided them into ten batches, each containing 1,200,000 bits. These batches were analyzed using the NIST SP 800-22 statistical test suite, with the outcomes summarized in Table 4, where the P-value is abbreviated as P-val. The results in Table 4 demonstrate that the eight generated keys possess a high degree of randomness. Each P-value in the NIST test suite represents the probability that a perfect random number generator could produce a sequence exhibiting less randomness than the sequence being tested, based on the specific non-randomness criteria evaluated by the test. A P-value greater than 0.01 indicates that the sequence is considered random, while a P-value less than 0.01 suggests the sequence is non-random.
Table 4.
NIST SP 800-22 Statistical test suite results on keys.
| Statistical Tests |
P-val |
P-val |
P-val |
P-val |
P-val |
P-val |
P-val |
P-val |
Range | Result |
|---|---|---|---|---|---|---|---|---|---|---|
| Frequency | 0.739918 | 0.534146 | 0.534146 | 0.534146 | 0.122325 | 0.534146 | 0.911413 | 0.739918 | 0.122325–0.911413 | ![]() |
| Block Frequency | 0.911413 | 0.739918 | 0.739918 | 0.911413 | 0.066882 | 0.739918 | 0.911413 | 0.213309 | 0.066882–0.911413 | ![]() |
| Cumulative Sums : Forward | 0.911413 | 0.350485 | 0.350485 | 0.122325 | 0.534146 | 0.534146 | 0.350485 | 0.739918 | 0.122325–0.911413 | ![]() |
| Cumulative Sums : Reverse | 0.739918 | 0.122325 | 0.066882 | 0.350485 | 0.213309 | 0.066882 | 0.122325 | 0.534146 | 0.066882–0.739918 | ![]() |
| Non Overlapping Template Matching | 0.350485 | 0.213309 | 0.739918 | 0.739918 | 0.350485 | 0.739918 | 0.739918 | 0.066882 | 0.066882–0.739918 | ![]() |
| Overlapping Template Matching | 0.911413 | 0.066882 | 0.911413 | 0.739918 | 0.739918 | 0.350485 | 0.739918 | 0.122325 | 0.066882–0.911413 | ![]() |
| Longest Run | 0.066882 | 0.739918 | 0.534146 | 0.122325 | 0.122325 | 0.911413 | 0.066882 | 0.035174 | 0.035174–0.911413 | ![]() |
| Fast Fourier Transform | 0.911413 | 0.350485 | 0.350485 | 0.534146 | 0.350485 | 0.911413 | 0.350485 | 0.213309 | 0.213309–0.911413 | ![]() |
| Matrix Rank | 0.739918 | 0.534146 | 0.739918 | 0.911413 | 0.350485 | 0.350485 | 0.534146 | 0.911413 | 0.350485–0.911413 | ![]() |
| Serial 1 | 0.066882 | 0.534146 | 0.350485 | 0.534146 | 0.213309 | 0.534146 | 0.911413 | 0.739918 | 0.066882–0.911413 | ![]() |
| Serial 2 | 0.534146 | 0.350485 | 0.534146 | 0.534146 | 0.739918 | 0.534146 | 0.739918 | 0.017912 | 0.017912–0.739918 | ![]() |
| Universal | 0.739918 | 0.534146 | 0.911413 | 0.350485 | 0.534146 | 0.739918 | 0.911413 | 0.350485 | 0.350485–0.911413 | ![]() |
| Approximate Entropy | 0.213309 | 0.534146 | 0.911413 | 0.911413 | 0.213309 | 0.350485 | 0.534146 | 0.122325 | 0.122325–0.911413 | ![]() |
| Linear Complexity | 0.739918 | 0.739918 | 0.534146 | 0.739918 | 0.911413 | 0.739918 | 0.991468 | 0.213309 | 0.213309–0.911413 | ![]() |
Proposed encryption and decryption process
In this section, we discuss the proposed encryption and decryption algorithms in more detail, along with pseudocodes providing the detailed explanation necessary for implementation.
Proposed encryption process
We begin our encryption process by taking the user image, which we further need to divide into its three principal components, namely, the red, green, and blue channels. The steps are as follows:
- Step 1: VPD Three RGB image channels are used. Each channel is converted into a binary bitstream. These three bitstreams are then concatenated into a single-row form.
- Generating sequence
: A sequence called
is constructed such that the first term of
consists of the first three bits from the concatenated bitstream. The second term of
includes the next three bits from the bitstream. This pattern is continued, where each subsequent
term corresponds to the next set of 3 bits from the bitstream. - Generating sequence
: Now, create a new sequence called
. The first term of
remains the same as that of
. For the subsequent terms of
, the
term of
is obtained by applying the XOR operation between the
term of
and the
term of
. In other words,
for
2. After that, the bitstream
is divided into three equal parts, and the resulting three mixed bitstreams are converted into similar types of channels, such as red, green, and blue, and then are combined to obtain a partially encrypted image. This step is crucial, as the adversary can attack any channel to avoid a channel attack. Additionally, an adversary cannot bypass this step, even if they use an image with uniform pixel values.
Step 2: 3D pixel intershuffling After that, one can use a new secure “3D pixel intershuffling” method. In this method, three iterators, i, j, and k, are used, which iterate over the image in a looping manner. Consider a pixel at position
in the image, where one can replace this pixel with the pixel available at position (
mod 255,
mod 255,
mod 255). In this way, one can generate very secure and random 3D pixel intershuffling.Step 3: Zigzag XORing Zigzag XORing is applied after finishing Step 2. The process of zigzag XORing takes an image and a key
as parameters. It begins by transforming the key by taking its absolute value, multiplying it by
, and then flooring the result to obtain an integer. Next, it applies a modulo operation with 256 to ensure that the key remains within the range of [0,256]. After preparing the key, the function retrieves the size s of the input image I. It then iterates through each pixel of the image in a triple nested loop, traversing through its width, height, and channels. At each pixel, it performs a bitwise XOR operation between the pixel value I(i, j, k) and the transformed key, storing the result in the corresponding location of the output image.Step 4: Repeat Steps 1, 2, and 3 After zigzag XORing, repeat Step 1; Step 2 with keys
and
; Step 3 with key
.Step 5: Planet with key
Then, at a time t, find the distance between the Sun and Uranus (Ur), Mars (Ma), Venus (Ve), Neptune (Ne), Jupiter (Ju), Earth (Ea), Saturn (Sa), and Mercury (Me). These eight distances are arranged in ascending order, and binary values between 000 and 111 are assigned. Now, iterate the image over three channels by converting it to an 8-bit binary, taking the three values at the i index, concatenating them, and assigning them to their corresponding planet name. Take a triplet of bits from the proposed key
after scaling its absolute value by a factor of 256 and XOR with the binary counterpart of the corresponding planet name. After this, the bitstream is divided into three parts: the red, green, and blue channels. These binary values are converted back into the original pixel values and concatenated to obtain our final encrypted image.
Proposed decryption process
In this section, we focus on the decryption process of encrypted images to obtain the original image.
Step 1: Reverse the planet with key
For each set of bits representing a planet name for each channel of the encrypted image, the XOR operation is performed with the corresponding bits from key
. The bitstream is divided into three parts for the red, green, and blue channels. These binary values are converted back into the original pixel values and concatenated.- Step 2: Reverse VPD In this process, each color channel is transformed into a binary format. The binary streams from the red, green, and blue channels are combined into a single binary stream.
- Generate sequence
: Construct a sequence called
such that the first term of
consists of the first three bits of the bitstream, the second term consists of the next three bits, and this pattern continues throughout the bitstream. - Generate sequence
: Generate another sequence
as follows:
by taking
as the initial term. After that, the bitstream
is divided into three parts: the red, green, and blue channels and the bitstream is converted into an image.
- Step 3: Reverse zigzag XORing Consider the key
, transform it like
where the symbol ‘
’ is defined as the greatest integer of a function, and ‘
’ is defined as a modulus function. We take the size s of the input image I. It then iterates through each pixel of the image in a triple nested loop, traversing through its width, height, and channels. At each pixel, it performs a bitwise XOR operation between the pixel value I(i, j, k) and the transformed key
, and the result is stored in the corresponding location of the output image. Step 4: Reverse 3D pixel intershuffling Iterate over each pixel location and reverse the shuffling process using the original values of
,
, and
to relocate each pixel to its original position.Step 5: Reverse Steps 2, 3, and 4 Repeat Steps 2, 3, and 4, and again Step 2 to return the original image. The encryption process in more detail is described in algorithms 2 and 4. The decryption process is explained in detail via Algorithms 3 and 5.
Algorithm 2.
Functions used in the proposed encryption algorithm 4
Algorithm 3.
Reverse functions used in the proposed decryption algorithm 5
Algorithm 4.
Proposed medical image encryption scheme
Algorithm 5.
Proposed medical decryption scheme
Encryption and decryption results
The outputs of the proposed algorithm on medical images (multiple images, brain MRI images, skin cancer images, and chest X-ray images) and a test baboon image are shown in Fig. 10.
Fig. 10.
(a–c) Multiple images (original, encrypted, and decrypted), (d–f) brain MRI images (original, encrypted, and decrypted), (g–i) skin cancer images (original, encrypted, and decrypted), (j–l) chest X-ray images (original, encrypted, and decrypted), (m–o) Baboon images (original, encrypted, and decrypted).
Security analysis
In this section, a detailed performance and security analysis is presented to evaluate the robustness and effectiveness of the proposed encryption algorithm. The feasibility of the approach is demonstrated through simulation results, which highlight the algorithm’s ability to ensure randomness in the encrypted data. This randomness contributes significantly to the robustness of the encryption algorithm, even under adversarial attacks. Analyzing encrypted images involves examining the strength of their protective measures to ensure that they remain safe from unauthorized access or alterations.
Time complexity
The time complexity of the encryption algorithm for an image of size
, where
, is determined by the individual complexities of its steps. The algorithm begins by converting the image into its RGB channels and subsequently into a concatenated bitstream. This step involves unpacking bits with a complexity of
as there are three channels for RGB images. The next step constructs sequence
by reshaping the bitstream, which is performed in
. The XOR operation to create
uses a cumulative XOR computation, also taking
. The intershuffling step, which employs nested loops across the three dimensions of the image, has a complexity of
. The zigzag XOR operation iterates through all pixels, also contributing
. Final steps, such as repeated VPD operations and planet encryption (using a chaotic map), iterated over the entire bitstream and have complexities of
. Consequently, the overall encryption complexity is
. The decryption process mirrors the encryption steps, resulting in a similar time complexity of
. This analysis highlights the linear scalability of the proposed encryption and decryption processes with respect to the image size, ensuring computational efficiency for high-resolution images. Although encryption time is typically used to evaluate processing speed, various operating environments can affect it. The experimental simulations were conducted on an HPC server equipped with an AMD Ryzen 9 7950X 16-core processor with 4.5GHz, supported by 125 GB RAM. The proposed algorithm was implemented and tested using Python 3.11.7. The encryption and decryption execution times of the proposed algorithm for different-sized images are listed in Table 5.
Table 5.
Encryption time (in seconds) for images of different sizes.
| Image size | Encryption | Decryption |
|---|---|---|
![]() |
0.29 | 0.27 |
![]() |
0.35 | 0.36 |
![]() |
0.63 | 0.60 |
![]() |
1.77 | 1.80 |
To ensure a fair comparison of the run time of the proposed algorithm with existing methods, we have evaluated it on a personal computer with specifications similar to those reported in other existing algorithms. Specifically, the tests were conducted on a system equipped with an 11th Gen Intel(R) Core(TM) i5-1135G7 processor running at 2.4 GHz and 16 GB of RAM. The results are summarized in Table 6.
Table 6.
Average values of encryption time (in seconds) of image size 256 x 256 compared with existing methods.
Key space analysis
The key space of a secure encryption algorithm must be larger than
to resist brute-force attacks66. The keyspace of the proposed algorithm is primarily determined by the precision of the eight initial values generated for chaotic sequence generation. These values are derived from the 512-bit output of SHA-512, which consolidates the entropy from the sum of pixel intensities of RGB channel sums of a DCGAN-generated image, a timestamp, and a nonce. While these inputs add randomness to the hashing process, the SHA-512 ensures a uniform 512-bit output, making the raw input complexity irrelevant to the keyspace. The 512-bit output is divided into eight parts, each contributing an initial value with a precision of
. This precision leads to a total keyspace of approximately
. Further, the encryption step involves planet XORing as provided in Table 3, which has 40320 options for choosing the planet XORing table. Hence, the resulting keyspace becomes
This ensures a vast keyspace, resistant to brute-force attacks, and comparable or superior to many existing encryption schemes, as shown in Table 7
Table 7.
Key space of various image encryption algorithms.
Histogram analysis
Histogram analysis provides a visual summary of the data distribution in a dataset. It shows the frequency or occurrence of different values within a range, helping to understand patterns such as peaks, gaps, or outliers. Displaying these frequencies as bars simplifies grasping the data’s central tendencies, spread, and overall shape, aiding in tasks like identifying common values, detecting anomalies, or making decisions on the basis of the data’s distribution. It is a graphical tool that provides an easy-to-understand overview of how values are distributed within a dataset. The histogram analysis of multiple images, brain MRI images, chest X-ray images, and baboon images are shown in Figs. 11, 12, 13, and 14. One can easily observe from the histogram of encrypted images in Figs. 11 (b, e, h), 12 (b, e, h), 13 (b, e, h), and 14 (b, e, h) show a constant level for each different image. Hence, we can conclude that the proposed algorithm is free from histogram attacks.
Fig. 11.
Histogram analysis of multiple images: (a–c) red channel (original, encrypted, and decrypted), (d–f) green channel (original, encrypted, and decrypted), and (g–i) blue channel (original, encrypted, and decrypted).
Fig. 12.
Histogram analysis of brain MRI image: (a–c) red channel (original, encrypted, and decrypted), (d–f) green channel (original, encrypted, and decrypted), and (g–i) blue channel (original, encrypted, and decrypted).
Fig. 13.
Histogram analysis of a chest X-ray image: (a) red channel (original, encrypted, and decrypted), (d) green channel (original, encrypted, and decrypted), and (g) blue channel (original, encrypted, and decrypted).
Fig. 14.
Histogram analysis of baboon image: (a) red channel (original, encrypted, and decrypted), (d) green channel (original, encrypted, and decrypted), and (g) blue channel (original, encrypted, and decrypted).
Information entropy
One of the important measures is the entropy (E) of an image. It measures randomness or uncertainty in the source image, see reference for more details69. It is computed using the following formula provided in Eq. (2):
![]() |
2 |
where I and N denote the image and number of intensity levels in the image, and the term p(i) denotes the probability of occurrence of intensity level i in the image. To calculate p(i), one needs to count the frequency of each intensity level in the source image and divide it by the total number of pixels. The result of this formula represents the entropy in bits per pixel (bpp), indicating the average amount of information required to encode each pixel in the image. A higher entropy value suggests greater randomness or complexity within the image. The information entropy of the RGB components of different images (multiple images, brain MRI image, skin cancer image, chest X-ray image, baboon image) is shown in Table 8.
Table 8.
Entropy for different encrypted images and their RGB components.
| Test encrypted images | Entropy | ||
|---|---|---|---|
| R | G | B | |
| Multiple images | 7.9992 | 7.9993 | 7.9992 |
| Brain MRI | 7.9993 | 7.9993 | 7.9992 |
| Skin cancer | 7.9993 | 7.9992 | 7.9993 |
| Chest X-ray | 7.9994 | 7.9992 | 7.9993 |
| Baboon | 7.9993 | 7.9993 | 7.9993 |
Key sensitivity
The proposed keys are sensitive to
decimal places. For instance, we demonstrate this statement via one of the keys, key 4 (
). The decryption process failed to yield the correct image when a perturbation of
was applied to the initial condition of
. Notably, a reduction in the perturbation to
resulted successfully (
) in recovering the original image during decryption. This observation highlights the algorithm’s sensitivity to minute variations in the key. The nuanced behavior underscores the significance of precise key management for ensuring the reliability and security of the cryptographic system. Thorough documentation remains imperative for comprehensive security assessments and the ongoing maintenance of system integrity. Decryption using a very marginal change, i.e., adding “
” and “
” to the initial condition of
performed on the images (multiple images and brain MRI image) are shown in Figs. 15 and 16.
Fig. 15.

(a) Original multiple images, (b) Decrypted image with a small change in the initial condition of
, (c) Correctly decrypted image.
Fig. 16.

(a) Original brain MRI image, (b) Decrypted image with a small change in the initial condition of
, (c) Correctly decrypted image.
Add noise to the encrypted image
Adding noise to an encrypted image is a method to make the encrypted data more secure. It works by scrambling the encrypted image with random information (noise) to hide any patterns that might be present. Here, we use white Gaussian noise (0.05%) and salt & pepper noise (0.02%). When the image is decrypted, this added noise ensures that the final image is clear and accurate. This process helps protect sensitive images from certain attacks, increasing the resistance of the encrypted data to unwanted access or deciphering attempts. The encrypted images obtained by adding noise to the original images (multiple images, brain MRI image) and the decrypted images are shown in Figs. 17 and 18.
Fig. 17.
(a) Original multiple images: (b) encrypted multiple images with 0.05% Gaussian noise, (c) decrypted multiple images with 0.05% Gaussian noisy encrypted image, (d) encrypted multiple images with 0.02% salt and pepper noise, (e) decrypted multiple images with 0.02% salt and pepper noisy encrypted image.
Fig. 18.
(a) Original brain MRI image, (b) encrypted brain MRI image with 0.05% Gaussian noise, (c) decrypted brain MRI image with 0.05% Gaussian noisy encrypted image, (d) encrypted brain MRI image with 0.02% salt & pepper noise, (e) decrypted brain MRI with 0.02% salt & pepper noisy encrypted image.
Correlation analysis
Table 9 displays the correlation coefficient values calculated using equation (3) for both the original and encrypted images for horizontal, vertical, and diagonal directions. The results (both before and after encryption) are plotted and shown in Figs. 19-21. The following formula is used to determine the correlation coefficient:
![]() |
3 |
In this context,
and
denote the values of two adjacent pixels while
represents the total number of samples. The analysis demonstrates that the proposed encryption scheme effectively diminishes the correlation between adjacent pixels in the encrypted images, enhancing security and data confidentiality. The results show that although the encryption process disrupts the original pixel correlations, these correlations are accurately restored during decryption.
Table 9.
Correlation values of encrypted images for RGB components.
| Testing encrypted image | Horizontal | Vertical | Diagonal | ||||||
|---|---|---|---|---|---|---|---|---|---|
| R | G | B | R | G | B | R | G | B | |
| Multiple images | −0.0005 | 0.0093 | −0.0549 | 0.0161 | 0.0051 | 0.0181 | −0.0016 | −0.0387 | −0.0820 |
| Brain MRI | 0.0045 | −0.0314 | 0.0114 | 0.0197 | −0.0154 | −0.0014 | 0.0918 | 0.0185 | −0.0231 |
| Skin cancer | 0.0218 | 0.0026 | 0.0158 | 0.0041 | −0.0304 | −0.0110 | −0.0272 | −0.0241 | −0.0184 |
| Chest X-ray | 0.0375 | 0.0035 | −0.0128 | 0.0031 | −0.0066 | 0.0124 | 0.0412 | 0.0833 | −0.0345 |
| Baboon | 0.0128 | 0.0060 | 0.0036 | 0.0057 | 0.0171 | −0.0166 | −0.0783 | −0.0079 | 0.1031 |
Fig. 19.
Correlation analysis of the multiple images before and after encryption: (a–b) red channel horizontal (original, encrypted), (c–d) green channel horizontal (original, encrypted), (e–f) blue channel horizontal (original, encrypted), (g–h) red channel vertical (original, encrypted), (i–j) green channel vertical (original, encrypted), (k–l) blue channel vertical (original, encrypted), (m-n) red channel diagonal (original, encrypted), (o–p) green channel diagonal (original, encrypted), and (q–r) blue channel diagonal (original, encrypted).
Fig. 21.
Correlation analysis of the baboon image before and after encryption: (a–b) red channel horizontal (original, encrypted), (c–d) green channel horizontal (original, encrypted), (e–f) blue channel horizontal (original, encrypted), (g–h) red channel vertical (original, encrypted), (i–j) green channel vertical (original, encrypted), (k–l) blue channel vertical (original, encrypted), (m-n) red channel diagonal (original, encrypted), (o–p) green channel diagonal (original, encrypted), and (q–r) blue channel diagonal (original, encrypted).
Correlation plots of original, encrypted, and decrypted images of different images are shown in Figs. 19, 20, and 21.
Fig. 20.
Correlation analysis of the chest X-ray image before and after encryption: (a–b) red channel horizontal (original, encrypted), (c–d) green channel horizontal(original, encrypted), (e–f) blue channel horizontal (original, encrypted), (g–h) red channel vertical (original, encrypted), (i–j) green channel vertical (original, encrypted), (k–l) blue channel vertical (original, encrypted), (m–n) red channel diagonal (original, encrypted), (o–p) green channel diagonal (original, encrypted), and (q–r) blue channel diagonal (original, encrypted).
Robustness against cropping attacks
Protecting against cropping tricks means teaching programs to remain accurate even when some parts of an image are changed or removed. It shows the program many different image versions where parts are cut off. Figures 22 and 23 show cropping attacks for different images.
Fig. 22.
Multiple images: (a–b) four corner cropped (encrypted image, decrypted image), (c–d) center cropped (encrypted image, decrypted image), (e) red channel one corner cropped encrypted image, (f) green channel one corner cropped encrypted image, (g) blue channel one corner cropped encrypted image, and (h) one corner cropped decrypted image.
Fig. 23.
Brain MRI image: (a–b) four corner cropped (encrypted image, decrypted image), (c–d) center cropped (encrypted image, decrypted image), (e) red channel one corner cropped encrypted image, (f) green channel one corner cropped encrypted image, (g) blue channel one corner cropped encrypted image, and (h) one corner cropped decrypted image.
Robustness against adversarial attacks
The encryption step effectively obfuscates adversarial patterns, safeguarding sensitive medical images during transmission or storage. By encrypting adversarial images generated using FGSM70 and DeepFool71 methods, the algorithm disrupts the visibility of adversarial perturbations, rendering them inaccessible to potential attackers without the decryption key. Figure 24 comparing the original, adversarial, encrypted, and decrypted images demonstrates how the encryption process masks adversarial patterns, transforming them into indistinguishable encrypted forms while preserving data integrity for subsequent decryption. If adversarial perturbations remain visible in the encrypted output, attackers might identify these as vulnerabilities to exploit. The imperceptibility of encrypted adversarial images indicates that the encryption algorithm is robust, uniform, and resilient against data-dependent attacks. It shows that the algorithm not only secures the data but also disrupts the potential of adversarial patterns to undermine system integrity, providing an additional layer of security beyond traditional encryption methods.
Fig. 24.
Adversarial attack using Skin cancer images (a–c) Original image encrypted and decrypted, (d–f) Adversarial image encrypted and decrypted.
Chosen-plaintext attacks: all-black and all-white image analysis
Chosen-plaintext attacks (CPA) often exploit images with uniform pixel values, such as entirely black (0) or completely white (255), as these are among the most commonly used plaintexts for testing the predictability of encryption algorithms. These inputs give attackers a straightforward way to analyze the relationship between plaintext and ciphertext, aiming to infer patterns or weaknesses in the encryption process. Figure 25 illustrates the encryption and decryption results for all-black and all-white images. The encrypted version of the image exhibited no visible patterns or resemblance to the original. It ensures that attackers cannot derive useful information from the ciphertext to compromise the encryption keys, validating the algorithm’s robustness against predictability.
Fig. 25.
Plain, cipher, and decrypted images of all-black (a–c) and all-white (d–f) images.
MSE and PSNR
In image processing, the MSE is a statistic that is used to measure the difference between two images. It is calculated by equation (4) using double summation as follows:
![]() |
4 |
where the image dimensions are indicated by
and
, and
and
are the intensities of the pixel at row
and column
in the original and distorted (encrypted) images, respectively.
Another often-used metric to evaluate an image’s quality is the PSNR. It is calculated via equation (5) as follows:
![]() |
5 |
The maximum possible pixel value in this case is represented by
(e.g., 255 for an 8-bit image). The PSNR provides a logarithmic scale to show the relationship between the greatest potential power and the error power. It is measured in decibels (dB). The MSE and PSNR values for the RGB components of several images are displayed in Table 10.
Table 10.
Image quality metrics (MSE and PSNR) for different encrypted images and their RGB components.
| Testing encrypted image | MSE | PSNR (dB) | ||||
|---|---|---|---|---|---|---|
| R | G | B | R | G | B | |
| Multiple images | 4348.2332 | 3377.4158 | 3268.5114 | 11.7477 | 12.8450 | 12.9873 |
| Brain MRI | 5707.8850 | 4455.6269 | 3683.5867 | 10.5611 | 11.6417 | 12.4681 |
| Skin cancer | 4735.3033 | 2857.8787 | 2555.8665 | 11.3773 | 13.5704 | 14.0554 |
| Chest X-ray | 4166.3107 | 3860.5932 | 3887.8420 | 11.9333 | 12.2643 | 12.2337 |
| Baboon | 2885.7602 | 2521.0127 | 3091.4646 | 13.5282 | 14.1151 | 13.2292 |
NPCR and UACI
The amount of pixel modifications in an encrypted image that occur from a single pixel change in the original image is measured using a metric called the NPCR. Equation (6) formulates it as follows:
![]() |
6 |
where
are the encrypted images caused by a one-pixel change in the original image, and
is a binary indicator, taking a value of 1 otherwise, and 0 when
. The multiplication by 100 highlights the rate per pixel change, and the summation is applied over the image’s dimensions.
The UACI offers insight into the average intensity of discrepancies between two encrypted images. It is calculated by equation (7) in the following manner:
![]() |
7 |
In this context,
and
represent the pixel values in encrypted images before and after a single-pixel modification in the original image. The ideal values for NPCR and UACI for an ideally encrypted image are
and
, respectively. Table 11 shows that the NPCR and UACI values for the RGB components of different images are around the theoretical value and thus secure against differential attacks. Additionally, the last two rows show values are passing the thresholds for encrypted images of adversarial images from FGSM and DeepFool, hence the adversarial attacks are unsuccessful.
Table 11.
Image quality metrics (NPCR and UACI) for different encrypted images and their RGB components.
| Testing encrypted image | NPCR % | UACI % | ||||
|---|---|---|---|---|---|---|
| R | G | B | R | G | B | |
| Multiple images | 99.6025 | 99.5964 | 99.6384 | 33.4548 | 33.5090 | 33.4709 |
| Brain MRI | 99.6227 | 99.5770 | 99.6044 | 33.5441 | 33.4396 | 33.4387 |
| Skin cancer | 99.5880 | 99.6059 | 99.6075 | 33.4218 | 33.4246 | 33.5662 |
| Chest X-ray | 99.6113 | 99.5972 | 99.5869 | 33.4716 | 33.3856 | 33.3821 |
| Baboon | 99.5953 | 99.6056 | 99.6109 | 33.3729 | 33.5050 | 33.4605 |
| Lena | 99.7643 | 99.6365 | 99.7959 | 33.3643 | 33.4455 | 33.4514 |
| Pepper | 99.8612 | 99.6971 | 99.9468 | 33.4971 | 33.4638 | 33.4560 |
| Breast Carcinoma | 99.8390 | 99.7631 | 99.6113 | 33.4790 | 33.4971 | 33.4550 |
| Gliomia | 99.0463 | 99.4289 | 99.3301 | 33.5192 | 33.4928 | 33.5194 |
| Barbara | 99.7219 | 99.8459 | 99.7749 | 33.3863 | 33.4585 | 33.3626 |
| Average of 10 images | 99.6253 | 99.6354 | 99.6507 | 33.4511 | 33.4622 | 33.4563 |
| Average of 50 random images | 99.6228 | 99.6241 | 99.6231 | 33.4680 | 33.4535 | 33.4995 |
| Adversarial attack (FGSM) | 99.6215 | 99.6521 | 99.6338 | 33.5962 | 33.5786 | 33.5520 |
| Adversarial attack (DeepFool) | 99.6216 | 99.5850 | 99.5849 | 33.5435 | 33.4294 | 33.4823 |
Statistical test for the NPCR
Following the methodology outlined in72, we evaluate the effectiveness of the proposed algorithm using statistical tests for the number of pixels change rate (NPCR). Given two encrypted images
and
, each of size
, we define the hypotheses (
and
) at a significance level
for
in equation (8) as follows:
![]() |
8 |
where F denotes the highest pixel value in the original image.
As observed from Table 12, the
values for all medical images, including multiple images, brain MRI, skin cancer, chest X-ray, and baboon, exceed the
values for
, 0.01, and 0.001. Thus, we accept the null hypothesis (
). Consequently, the NPCR values indicate that the proposed algorithm is robust and reliable.
Table 12.
Statistical test for the NPCR.
| Testing encrypted-image | F = 255 | ||||
|---|---|---|---|---|---|
![]() |
![]() |
![]() |
![]() |
![]() |
|
| Numerical values | 99.6094 | 0.0122 | 99.5893 | 99.5810 | 99.5717 |
| Multiple images (99.6124) | ![]() |
![]() |
![]() |
||
| Brain MRI (99.6013) | ![]() |
![]() |
![]() |
||
| Skin cancer (99.6004) | ![]() |
![]() |
![]() |
||
| Chest X-ray (99.5984) | ![]() |
![]() |
![]() |
||
| Baboon (99.6039) | ![]() |
![]() |
![]() |
||
Statistical test for UACI
Similarly, based on72, we can demonstrate that the proposed algorithm is effective via statistical tests for the unified average changing intensity (UACI). Assuming that we have two encrypted images
and
, each of size
, we define the hypotheses (
and
) at a significance level
for
in equation (9) are as follows:
![]() |
9 |
where F is the highest pixel value in the original image.
From Table 13, we observe that the
values for all medical images, including multiple images, brain MRI, skin cancer image, chest X-ray image, and baboon image, fall within the interval
for
, 0.01, and 0.001. Thus, we accept the null hypothesis (
). Therefore, the UACI values indicate that the proposed algorithm is effective and reliable.
Table 13.
Statistical test for UACI.
| Testing encrypted-image | F = 255 | ||||
|---|---|---|---|---|---|
![]() |
![]() |
![]() |
![]() |
![]() |
|
| Numerical values | 33.4635 | 0.0462 | 33.3730 33.5541 | 33.3445 33.5826 | 33.3115 33.6156 |
| Multiple images (33.4782) | ![]() |
![]() |
![]() |
||
| Brain MRI (33.4741) | ![]() |
![]() |
![]() |
||
| Skin cancer (33.4708) | ![]() |
![]() |
![]() |
||
| Chest X-ray (33.4131) | ![]() |
![]() |
![]() |
||
| Baboon (33.4461) | ![]() |
![]() |
![]() |
||
NIST SP 800-22 Statistical test suite results
Here, we generated 18,000,000 bits and divided them into ten parts, each consisting of 1,800,000 bits. These batches were then analyzed via the NIST SP 800-22 statistical test suite. The results are presented in Table 14 below:
Table 14.
The NIST SP 800-22 statistical test suite results on generated outputs of encrypted images.
| Statistical Tests | P-Values | Result |
|---|---|---|
| Frequency | 0.739918 | ![]() |
| Block Frequency | 0.122325 | ![]() |
| Cumulative Sums : Forward | 0.350485 | ![]() |
| Cumulative Sums : Reverse | 0.213309 | ![]() |
| Non Overlapping Template Matching | 0.350485 | ![]() |
| Overlapping Template Matching | 0.122325 | ![]() |
| Longest Run | 0.122325 | ![]() |
| Fast Fourier Transform | 0.739918 | ![]() |
| Matrix Rank | 0.911413 | ![]() |
| Serial 1 | 0.350485 | ![]() |
| Serial 2 | 0.739918 | ![]() |
| Universal | 0.991468 | ![]() |
| Approximate Entropy | 0.534146 | ![]() |
| Linear Complexity | 0.534146 | ![]() |
Comparative analysis
In the realm of security analysis, it is imperative to compare methodologies and outcomes across various research papers to discern overarching trends, identify best practices, and pinpoint gaps in current knowledge. A robust comparative analysis typically focuses on contrasting the effectiveness, efficiency, and scalability of different security solutions or frameworks. Table 15 shows some comparisons with other medical image encryption papers.
Table 15.
Comparison with other existing encryption algorithms.
| Medical Image | Algo | Channel | NPCR | UPCR | IE | Correlation | |||
|---|---|---|---|---|---|---|---|---|---|
| Horizontal | Vertical | Diagonal | Average correlation | ||||||
| Baboon | 73 | R | 99.5900 | 33.4600 | 7.9998 | −0.0022 | 0.0022 | −0.0019 | −0.00059 |
| G | 99.6000 | 33.4400 | 7.9998 | −0.00017 | −0.0004 | −0.0005 | −0.0012 | ||
| B | 99.6100 | 33.4700 | 7.9998 | −0.0006 | −0.0054 | −0.0032 | −0.0012 | ||
| 74 | R | 99.6536 | 33.4753 | 7.9972 | - | - | - | −0.0008 | |
| G | 99.6078 | 33.5090 | 7.9972 | - | - | - | 0.0002 | ||
| B | 99.6520 | 33.4176 | 7.9972 | - | - | - | −0.0006 | ||
| 75 | R | 99.6140 | 33.4843 | 7.9970 | - | - | - | 0.0013 | |
| G | 99.6073 | 33.4690 | 7.9978 | - | - | - | 0.0025 | ||
| B | 99.6292 | 33.4965 | 7.9987 | - | - | - | 0.0010 | ||
| 76 | R | 99.6118 | 33.4692 | 7.9898 | - | - | - | - | |
| G | 99.6084 | 33.4806 | 7.9897 | - | - | - | - | ||
| B | 99.6156 | 33.5040 | 7.9895 | - | - | - | - | ||
| Proposed | R | 99.5953 | 33.3729 | 7.9993 | 0.0128 | 0.0060 | 0.0036 | −0.01994 | |
| G | 99.6056 | 33.5050 | 7.9993 | 0.0057 | 0.0171 | −0.0166 | 0.0050 | ||
| B | 99.6109 | 33.4605 | 7.9993 | −0.0783 | −0.0079 | 0.1031 | 0.035 | ||
| Brain MRI | 77 | R | 99.62959 | 47.97603 | 7.99928 | 0.00146 | −0.00118 | −0.00015 | - |
| G | 99.62539 | 46.42358 | 7.99922 | 0.00072 | 0.00145 | −0.00101 | - | ||
| B | 99.61779 | 46.68882 | 7.99937 | 0.00242 | 0.00085 | −0.00552 | - | ||
| 78 | R | 99.5971 | 33.4136 | 7.99741 | −0.00253 | −0.00270 | 0.00211 | - | |
| G | 99.5819 | 33.4723 | 7.99756 | −0.00372 | −0.00147 | −0.00031 | - | ||
| B | 99.6032 | 33.4574 | 7.99711 | −0.00834 | 0.01012 | −0.00046 | - | ||
| Proposed | R | 99.6227 | 33.5441 | 7.9993 | 0.0045 | 0.0197 | 0.0918 | - | |
| G | 99.5770 | 33.4396 | 7.9993 | −0.0314 | −0.0154 | 0.0185 | - | ||
| B | 99.6044 | 33.4387 | 7.9992 | −0.0114 | −0.0014 | −0.0231 | - | ||
Conclusion
This paper presents a novel encryption technique that integrates DCGAN associated with the VPD approach to enhance the protection of medical images. The method uses a DL framework to generate a decoy image, which forms the basis for generating encryption keys using a timestamp, nonce, and 1-DEC map. The experimental results validate the efficacy of the approach in protecting medical images from various security threats, including unauthorized access, tampering, and adversarial attacks. Its computational complexity,
, ensures scalability for encrypting images of varying sizes, striking an effective balance between security and time efficiency as can be seen in tables 5 and 6. The large keyspace of
, derived from chaotic sequences with a precision of
and VPD XORing table, provides strong resistance against brute-force attacks and is comparable to other methods as shown in 7. The randomness of the keys and encrypted images are demonstrated through the NIST SP 800-22 statistical test suite provided in Tables 4 and 14, respectively. The robustness against key sensitivity, noise, and cropping attacks is shown in Figures 15-18, and 22-23. Security analysis results are shown (such as histogram plots in Figures 11-14 and correlation plots in Figures 19-21). The IE (
), correlation coefficient (
), MSE(
), PSNR (
), NPCR (
), and UACI (
) underscore the high security and reliability of the encrypted images, as shown in Tables 8-11. Furthermore, the statistical NPCR and UACI are calculated in Tables 12 and 13, respectively. The proposed algorithm is also compared with existing algorithms, and compared values are provided in Table 15. The data presented in Tables 4-15 highlight the practical viability and superior security of the proposed approach.
In the future, expanding the framework to video encryption presents another promising avenue. Recent advances in Selective Video Encryption Algorithms (SVEA)30 and temporal action segmentation79 highlight methods for optimizing security and efficiency by focusing on critical video frames or segments. These approaches, which use techniques like chaotic maps for one-step encryption, reduce computational overhead while maintaining high security, making them ideal for real-time applications in surveillance and streaming services. Moreover, selective video encryption principles could be adapted to optimize performance by prioritizing regions of interest within a video frame. Hybrid chaotic systems and advanced neural network architectures could be employed to provide security across diverse multimedia datasets80. These advancements, coupled with scalable algorithms like CP or Tucker decompositions81, MACH82, etc., for real-time tensor decompositions, provide a pathway to enhance the adaptability, security, and computational efficiency of the encryption framework against sophisticated attacks. Several GAN architectures, including Conditional GANs (CGANs) and Wasserstein GANs (WGANs), or Hybrid-GANs can also be adapted for specific data types, including images, music, and tabular data83.
Acknowledgements
The authors express their sincere gratitude to the reviewers and the editor for their valuable feedback and constructive suggestions, which have significantly enhanced the quality of this work. The authors also gratefully acknowledge the computing time provided on the high-performance computing facility, Sharanga, at the Birla Institute of Technology and Science - Pilani, Hyderabad Campus.
Funding
Open access funding provided by Birla Institute of Technology and Science.
Data Availability
The datasets (i.e., cancer images) used for key generation in this work are publicly available on Kaggle (under the terms specified by the Kaggle platform). It can be accessed via the link provided in "Key generation procedure" .
Declarations
Competing Interests
The authors declare that they do not have any competing financial or nonfinancial interests related to this work.
Footnotes
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Liu, H., Teng, L., Zhang, Y., Si, R. & Liu, P. Mutil-medical image encryption by a new spatiotemporal chaos model and DNA new computing for information security. Expert Systems with Applications235, 121090. 10.1016/j.eswa.2023.121090 (2024). [Google Scholar]
- 2.Dhingra, D. & Dua, M. Medical video encryption using novel 2D cosine-sine map and dynamic DNA coding. Medical & Biological Engineering & Computing62, 237–255 (2024). [DOI] [PubMed] [Google Scholar]
- 3.Bencherqui, A. et al. Optimal algorithm for color medical encryption and compression images based on DNA coding and a hyperchaotic system in the moments. Engineering Science and Technology, an International Journal50, 101612 (2024). [Google Scholar]
- 4.Sarathe, S. & Shrivastava, C. G. D. R. Medical image encryption using DNA sequencing and affine transform. International Journal of Trendy Research in Engineering and Technology7, 2582–0958 (2023). [Google Scholar]
- 5.Jamal, S. S. et al. Region of interest-based medical image encryption technique based on chaotic S-boxes. Expert Systems with Applications238, 122030. 10.1016/j.eswa.2023.122030 (2024). [Google Scholar]
- 6.Razaq, A., Maghrabi, L. A., Ahmad, M., Aslam, F. & Feng, W. Fuzzy logic-based substitution-box for robust medical image encryption in telemedicine. IEEE Access[SPACE]10.1109/ACCESS.2024.3351794 (2024). [Google Scholar]
- 7.Ahmed, S. M., MA Elkamchouchi, H., Elfahar, A., El-Shafai, W. & Mohamed, A. G. A hybrid medical image cryptosystem based on 4D-hyperchaotic S-boxes and logistic maps. Multimedia Tools and Applications83, 8837–8865. 10.1007/s11042-023-15925-6 (2024). [Google Scholar]
- 8.Güvenoğlu, E. & Tunalı, V. ZigZag transform with Durstenfeld shuffle for fast and secure image encryption. Connection Science35, 2162000 (2023). [Google Scholar]
- 9.Kumar, C. M., Vidhya, R. & Brindha, M. An efficient chaos based image encryption algorithm using enhanced thorp shuffle and chaotic convolution function. Applied Intelligence52, 2556–2585 (2022). [Google Scholar]
- 10.Levcopoulos, C. & Petersson, O. Sorting shuffled monotone sequences. Information and Computation112, 37–50 (1994). [Google Scholar]
- 11.Agrawal, P., Sahu, S. & Choudhary, A. A new method of MPEG video encryption using frame shuffling, 10.1049/iet-ipr.2018.5250 (2018). [Google Scholar]
- 12.Reyad, O., Mansour, H. . M., Zanaty, E. . A. & Heshmat, M. Medical image privacy enhancement using key extension of advanced encryption standard. Sohag Journal of Sciences9, 211–216. 10.21608/SJSCI.2024.250176.1154 (2024). [Google Scholar]
- 13.Hadj Brahim, A., Ali Pacha, A. & Hadj Said, N. An image encryption scheme based on a modified AES algorithm by using a variable S-box. Journal of Optics53, 1170–1185. 10.1007/s12596-023-01232-8 (2024). [Google Scholar]
- 14.Alarood, A. A., Faheem, M., Al-Khasawneh, M. A., Alzahrani, A. I. & Alshdadi, A. A. Secure medical image transmission using deep neural network in e-health applications. Healthcare Technology Letters10, 87–98 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Panwar, K. et al. Encipher GAN: An end-to-end color image encryption system using a deep generative model. Systems11, 36 (2023). [Google Scholar]
- 16.Rupa, C., Harshitha, M., Srivastava, G., Gadekallu, T. R. & Maddikunta, P. K. R. Securing multimedia using a deep learning based chaotic logistic map. IEEE Journal of Biomedical and Health Informatics27, 1154–1162 (2022). [DOI] [PubMed] [Google Scholar]
- 17.Fang, P., Liu, H. & Wu, C. A novel chaotic block image encryption algorithm based on deep convolutional generative adversarial networks. IEEE Access9, 18497–18517 (2020). [Google Scholar]
- 18.Rao, K. N., Jayasree, P., Krishna, C. V. M., Prasanth, K. S. & Reddy, C. S. Image anonymization using deep convolutional generative adversarial network. In Journal of Physics: Conference Series, vol. 2089, 012012 (IOP Publishing, 2021).
- 19.Krishna, A. A., Arikutharam, V., Ramnan, K. V., Bharathi, H. & Chandar, T. Dynamic image encryption using neural networks for medical images. In 2022 IEEE IAS Global Conference on Emerging Technologies (GlobConET), 739–745 (IEEE, 2022).
- 20.Kumar, P., Rahman, M., Namasudra, S. & Moparthi, N. R. Enhancing security of medical images using deep learning, chaotic map, and hash table. Mobile Networks and Applications 1–15 (2023).
- 21.Cai, C., Wang, Y., Cao, Y., Sun, B. & Mou, J. Multiple remote sensing image encryption scheme based on saliency extraction and magic cube circular motion. Applied Intelligence 1–17, 10.1007/s10489-024-05447-5 (2024).
- 22.Zhu, X., Lai, Z., Liang, Y., Xiong, J. & Wu, J. Generative high-capacity image hiding based on residual CNN in wavelet domain. Applied Soft Computing115, 108170. 10.1016/j.asoc.2021.108170 (2022). [Google Scholar]
- 23.Narayanan, K. L. & Naresh, R. Privacy-preserving dual interactive Wasserstein generative adversarial network for cloud-based road condition monitoring in VANETs. Applied Soft Computing154, 111367. 10.1016/j.asoc.2024.111367 (2024). [Google Scholar]
- 24.Lakhan, A., Grønli, T.-M., Muhammad, G. & Tiwari, P. EDCNNS: Federated learning enabled evolutionary deep convolutional neural network for Alzheimer disease detection. Applied Soft Computing147, 110804. 10.1016/j.asoc.2023.110804 (2023). [Google Scholar]
- 25.Kuznetsov, N. et al. Hidden attractors in Chua circuit: mathematical theory meets physical experiments. Nonlinear Dynamics111, 5859–5887. 10.1007/s11071-022-08078-y (2022). [Google Scholar]
- 26.Liu, L., Miao, S., Hu, H. & Deng, Y. Pseudorandom bit generator based on non-stationary logistic maps. IET Information Security10, 87–94 (2016). [Google Scholar]
- 27.Wang, M. et al. A new 2D-HELS hyperchaotic map and its application on image encryption using RNA operation and dynamic confusion. Chaos, Solitons & Fractals183, 114959. 10.1016/j.chaos.2024.114959 (2024). [Google Scholar]
- 28.Masood, F. et al. A lightweight chaos-based medical image encryption scheme using random shuffling and XOR operations. Wirel. Pers. Commun.127, 1405–1432. 10.1007/s11277-021-08584-z (2022). [Google Scholar]
- 29.Gao, S. et al. Design, hardware implementation, and application in video encryption of the 2-D memristive cubic map. IEEE Internet of Things Journal11, 21807–21815. 10.1109/JIOT.2024.3376572 (2024). [Google Scholar]
- 30.Gao, S. et al. Development of a video encryption algorithm for critical areas using 2D extended schaffer function map and neural networks. Applied Mathematical Modelling134, 10.1016/j.apm.2024.06.016 (2024).
- 31.Rehman, S. U. et al. Optimization of CNN through novel training strategy for visual classification problems. Entropy20, 10.3390/e20040290 (2018). [DOI] [PMC free article] [PubMed]
- 32.Rehman et al., S. U. Unsupervised pre-trained filter learning approach for efficient convolution neural network. Neurocomputing365, 171–190. 10.1016/j.neucom.2019.06.084 (2019). [Google Scholar]
- 33.Rehman, S. U., Tu, S., Huang, Y. & Yang, Z. Face recognition: A novel un-supervised convolutional neural network method. In 2016 IEEE International Conference of Online Analysis and Computing Science (ICOACS), 139–144, 10.1109/ICOACS.2016.7563066 (2016).
- 34.Kuznetsov, O., Zakharov, D. & Frontoni, E. Deep learning-based biometric cryptographic key generation with post-quantum security. Multimedia Tools and Applications83, 1–30. 10.1007/s11042-023-17714-7 (2023). [Google Scholar]
- 35.Rehman, S. U. et al.Deep Learning Models for Intelligent Healthcare: Implementation and Challenges, 214–225 (2021).
- 36.Ding, Y. et al. DeepEDN: A deep-learning-based image encryption and decryption network for internet of medical things. IEEE Internet of Things Journal8, 1504–1518 (2020). [Google Scholar]
- 37.Wang, C. & Zhang, Y. A novel image encryption algorithm with deep neural network. Signal Processing196, 108536. 10.1016/j.sigpro.2022.108536 (2022). [Google Scholar]
- 38.Goodfellow, I. J. et al. Generative adversarial networks1406, 2661 (2014). [Google Scholar]
- 39.Mirza, M. & Osindero, S. Conditional generative adversarial nets. CoRRabs/1411.1784, 10.48550/arXiv.1411.1784 (2014). 1411.1784.
- 40.Salimans, T. et al. Improved techniques for training GANs. CoRRabs/1606.03498 (2016). 1606.03498.
- 41.Isola, P., Zhu, J.-Y., Zhou, T. & Efros, A. A. Image-to-image translation with conditional adversarial networks1611, 07004 (2018). [Google Scholar]
- 42.Zhu, J., Park, T., Isola, P. & Efros, A. A. Unpaired image-to-image translation using cycle-consistent adversarial networks. CoRRabs/1703.10593 (2017). 1703.10593.
- 43.Karras, T., Aila, T., Laine, S. & Lehtinen, J. Progressive growing of GANs for improved quality, stability, and variation (2018). 1710.10196.
- 44.Brock, A., Donahue, J. & Simonyan, K. Large scale GAN training for high fidelity natural image synthesis. CoRRabs/1809.11096 (2018). 1809.11096.
- 45.Radford, A., Metz, L. & Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks1511, 06434 (2016). [Google Scholar]
- 46.Liu, B., Lv, J., Fan, X., Luo, J. & Zou, T. Application of an improved DCGAN for image generation. Mobile Information Systems2022, 10.1155/2022/9005552 (2022).
- 47.Panwar, K., Kukreja, S., Singh, A. & Singh, K. Towards deep learning for efficient image encryption. Procedia Computer Science218, 644–650. 10.1016/j.procs.2023.01.046 (2023). [Google Scholar]
- 48.Zheng, Z. et al. EncryptGAN: Image steganography with domain transform, 10.48550/arXiv.1905.11582 (2019).
- 49.Ding, Y. et al. DeepKeyGen: a deep learning-based stream cipher generator for medical image encryption and decryption. IEEE Transactions on Neural Networks and Learning Systems33, 4915–4929 (2021). [DOI] [PubMed] [Google Scholar]
- 50.Springenberg, J. T., Dosovitskiy, A., Brox, T. & Riedmiller, M. Striving for simplicity: The all convolutional net1412, 6806 (2015). [Google Scholar]
- 51.Ioffe, S. & Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. CoRRabs/1502.03167 (2015). 1502.03167.
- 52.Nair, V. & Hinton, G. E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning, ICML’10, 807-814 (Omnipress, Madison, WI, USA, 2010).
- 53.Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B. & Hochreiter, S. GANs trained by a two time-scale update rule converge to a local nash equilibrium1706, 08500 (2018). [Google Scholar]
- 54.Long, G. et al. Exploiting one-dimensional exponential Chebyshev chaotic map and matching embedding for visually meaningful image encryption. Chaos, Solitons & Fractals176, 114111 (2023). [Google Scholar]
- 55.Talhaoui, M. Z., Wang, X. & Midoun, M. A. A new one-dimensional cosine polynomial chaotic map and its use in image encryption. Vis. Comput.37, 541–551. 10.1007/s00371-020-01822-8 (2021). [Google Scholar]
- 56.Mansouri, A. & Wang, X. A novel one-dimensional sine powered chaotic map and its application in a new image encryption scheme. Inf. Sci520, 46–62 (2020). [Google Scholar]
- 57.Zhuang, Z., Zhuang, Z. & Wang, T. Medical image encryption algorithm based on a new five-dimensional multi-band multi-wing chaotic system and QR decomposition. Scientific Reports14, 402 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Sun, S. A new image encryption scheme based on 6D hyperchaotic system and random signal insertion. IEEE Access (2023).
- 59.John, S. & Kumar, S. 6D hyperchaotic encryption model for ensuring security to 3D printed models and medical images. Journal of Image and Graphics12, 117–126 (2024). [Google Scholar]
- 60.Niu, Y., Zhou, H. & Zhang, X. Image encryption scheme based on improved four-dimensional chaotic system and evolutionary operators. Scientific Reports14, 7033 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Kumar, M., Mohapatra, R. N., Agarwal, S., Sathish, G. & Raw, S. A new RGB image encryption using generalized Vigenére-type table over symmetric group associated with virtual planet domain. Multimedia Tools and Applications78, 10227–10263. 10.1007/s11042-018-6586-0 (2019). [Google Scholar]
- 62.Iqbal, N. et al. DNA strands level scrambling based color image encryption scheme. IEEE Access8, 178167–178182. 10.1109/ACCESS.2020.3025241 (2020). [Google Scholar]
- 63.Hua, Z. & Zhou, Y. Image encryption using 2D logistic-adjusted-sine map. Information Sciences339, 237–253. 10.1016/j.ins.2016.01.017 (2016). [Google Scholar]
- 64.Wang, X.-Y., Zhang, Y.-Q. & Bao, X.-M. A novel chaotic image encryption scheme using DNA sequence operations. Optics and Lasers in Engineering73, 53–61. 10.1016/j.optlaseng.2015.03.022 (2015). [Google Scholar]
- 65.Yan, X., Wang, X. & Xian, Y. Chaotic image encryption algorithm based on arithmetic sequence scrambling model and DNA encoding operation. Multimedia Tools and Applications80, 1–35. 10.1007/s11042-020-10218-8 (2021). [Google Scholar]
- 66.Ge, B., Chen, X., Chen, G. & Shen, Z. Secure and fast image encryption algorithm using hyper-chaos-based key generator and vector operation. IEEE AccessPP, 1–1, 10.1109/ACCESS.2021.3118377 (2021).
- 67.Chai, X., Gan, Z. & Zhang, M. A fast chaos-based image encryption scheme with a novel plain image-related swapping block permutation and block diffusion. Multimedia Tools and Applications76, 10.1007/s11042-016-3858-4 (2017).
- 68.Liu, L., Lei, Y. & Wang, D. A fast chaotic image encryption scheme with simultaneous permutation-diffusion operation. IEEE Access8, 27361–27374. 10.1109/ACCESS.2020.2971759 (2020). [Google Scholar]
- 69.Shannon, C. E. A mathematical theory of communication. The Bell system technical journal27, 379–423 (1948). [Google Scholar]
- 70.Goodfellow, I. J., Shlens, J. & Szegedy, C. Explaining and harnessing adversarial examples1412, 6572 (2015). [Google Scholar]
- 71.Moosavi-Dezfooli, S.-M., Fawzi, A. & Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. 2574–2582, 10.1109/CVPR.2016.282 (2016).
- 72.Wu, Y., Noonan, J. P., Agaian, S. et al. NPCR and UACI randomness tests for image encryption. Cyber journals: multidisciplinary journals in science and technology, Journal of Selected Areas in Telecommunications (JSAT)1, 31–38 (2011). [Google Scholar]
- 73.Yahi, A., Bekkouche, T., Daachi, M. E. H. & Diffellah, N. A color image encryption scheme based on 1D cubic map. Optik249, 168290 (2022). [Google Scholar]
- 74.Yaghouti Niyat, A., Moattar, M. . H. & Niazi Torshiz, M. Color image encryption based on hybrid hyper-chaotic system and cellular automata. Optics and Lasers in Engineering90, 225–237. 10.1016/j.optlaseng.2016.10.019 (2017). [Google Scholar]
- 75.Mohamed, H. G., ElKamchouchi, D. H. & Moussa, K. H. A novel color image encryption algorithm based on hyperchaotic maps and mitochondrial DNA sequences. Entropy22, 158 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Wu, X., Wang, K., Wang, X., Kan, H. & Kurths, J. Color image DNA encryption using NCA map-based CML and one-time keys. Signal Processing148, 272–287. 10.1016/j.sigpro.2018.02.028 (2018). [Google Scholar]
- 77.Parvees, M. Y. M. & Vijayakumar, T. Medical image cryptosystem using improved quadratic congruential generator and logistic map. Measurement: Sensors24, 100502, 10.1016/j.measen.2022.100502 (2022).
- 78.Deb, S. & Bhuyan, B. Chaos-based medical image encryption scheme using special nonlinear filtering function based LFSR. Multimedia Tools and Applications80, 19803–19826. 10.1007/s11042-020-10308-7 (2021). [Google Scholar]
- 79.Gao, S. et al. Temporal action segmentation for video encryption. Chaos, Solitons & Fractals183, 114958. 10.1016/j.chaos.2024.114958 (2024). [Google Scholar]
- 80.SaberiKamarposhti, M., Ghorbani, A. & Yadollahi, M. A comprehensive survey on image encryption: Taxonomy, challenges, and future directions. Chaos, Solitons & Fractals178, 114361. 10.1016/j.chaos.2023.114361 (2024). [Google Scholar]
- 81.Grasedyck, L., Kressner, D. & Tobler, C. A literature survey of low-rank tensor approximation techniques1302, 7121 (2013). [Google Scholar]
- 82.Tsourakakis, C. E. MACH: Fast randomized tensor decompositions0909, 4969 (2009). [Google Scholar]
- 83.Figueira, A. & Vaz, B. Survey on synthetic data generation, evaluation methods and GANs. Mathematics (2022).
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The datasets (i.e., cancer images) used for key generation in this work are publicly available on Kaggle (under the terms specified by the Kaggle platform). It can be accessed via the link provided in "Key generation procedure" .


































































































































