Computers 14 00055
Computers 14 00055
approaches to known threats, which may be insufficient in addressing the advanced tactics
employed by modern attackers [4].
This underscores the need for innovative solutions, such as AI-driven systems,
which can ensure data privacy while enhancing proactive threat detection capabilities [5].
The potential of artificial intelligence, particularly Generative Adversarial Networks (GANs)
and Variational Autoencoders (VAEs), offers a transformative approach to addressing these
challenges. These generative models excel at creating synthetic data that closely resem-
ble real-world datasets while preserving privacy, making them ideal for cybersecurity
applications [6].
GANs, introduced by Ian Goodfellow in 2014, consist of two neural networks—a
generator and a discriminator—trained in a competitive setting to produce highly realistic
data samples [7]. Similarly, VAEs use probabilistic modeling to encode and decode data,
enabling smooth interpolation and exploration of latent features [8].
This study introduces a Generative AI-Enhanced Cybersecurity Framework that lever-
ages GANs, VAEs, and traditional anomaly detection methods to address current data pri-
vacy and cybersecurity limitations. By integrating these advanced technologies, the frame-
work enables enterprises to perform the following:
• Generate synthetic datasets that simulate real-world scenarios for robust anomaly
detection training.
• Implement differential privacy, encryption, and data masking to ensure compliance
with regulatory standards.
• Detect and mitigate emerging threats proactively through real-time monitoring and
adaptive anomaly detection algorithms.
The proposed framework is a conceptual model developed to explore the potential of
GANs and VAEs in generating synthetic datasets and enhancing anomaly detection [9,10].
Experimental evaluations were performed using controlled simulations and synthetic data
to assess the framework’s performance metrics, such as recall and precision. While these
experiments demonstrate the feasibility and potential of the framework, its application in
real-world scenarios requires further validation and scalability testing.
The subsequent sections provide a detailed description of the framework’s components
and step-by-step guidance for implementation.
2. Literature Review
In traditional cybersecurity contexts, the National Institute of Standards and Technol-
ogy (NIST) Cybersecurity Framework, ISO/IEC 27001, and controls provide structured
approaches to managing cybersecurity risks [11,12]. These frameworks encompass guide-
lines for risk assessment, access control, data encryption, threat detection, incident response,
and continuous monitoring.
approach strengthens data privacy management and enhances robustness against emerging
threats as shown in the Table 1.
3. Proposed Framework
The provided framework, as shown in Figure 2, illustrates the end-to-end flow of
the proposed framework, from data collection to real-time anomaly detection and privacy
management. The visual highlights scalability and adaptability across domains and in-
tegrates advanced technologies to enhance cybersecurity by leveraging data synthesis,
anomaly detection, privacy management, and real-time monitoring. The framework pro-
tects systems from potential cyber threats by utilizing machine learning (ML) and deep
learning techniques for anomaly detection, with privacy-preserving measures and real-time
analytics. It outlines the flow from raw data sources to actionable insights, emphasizing
accuracy, efficiency, and privacy.
Generative Adversarial Networks (GANs) are a class of artificial intelligence models that
generate synthetic data resembling real-world data. GANs consist of two main components:
• Generator (G): Takes noise as input, along with additional data (e.g., text embed-
dings, latent codes, mutual information, labels, or data augmentation). It learns to
generate synthetic data by capturing underlying latent features such as cause-effect
relationships and distributed representations [36].
• Discriminator (D): A classifier that evaluates whether the generated data are real or fake.
It receives both real training data and synthetic samples generated by the generator.
• Input Noise: Random noise input to the generator, used to create variability and
diversity in the synthetic samples.
• Additional Input: Extra input to the generator, such as:
– Text embeddings: Representations of input data in a vectorized format.
– Latent codes: Encoded features that guide the generator to create diverse outputs.
– Mutual information: Measures correlations between generated features to im-
prove representation learning.
– Conditional labels: Labels that guide the generator to create class-specific outputs.
– Augmentation: Modified inputs to enhance the diversity of synthetic data.
Computers 2025, 14, 55 5 of 18
• Feedback Loop:
– The discriminator sends feedback by classifying generated data as real or fake.
– Both the generator and discriminator are updated iteratively, improving their
respective performances.
• Sample Evaluation (Decision Diamond Symbol): After generating synthetic data, an
evaluation step checks whether the samples are realistic.
– If the samples fail the evaluation, they are discarded (indicated by the trash
can symbol).
– If the samples pass, they are used to train a supervised learning model (repre-
sented by a magnifying glass and document symbol).
• Real vs. Fake Classification: The discriminator outputs whether a sample is real or
fake using a decision flow symbol that branches based on the outcome.
• Objective of GANs: To train the generator to produce high-quality synthetic data that
the discriminator finds increasingly difficult to distinguish from real data. Through
the adversarial training process, both the generator and discriminator continuously
improve their tasks [37].
Domain Application
Retail Generating customer behavior patterns for better service analysis.
Healthcare Creating medical images for AI-assisted diagnostics.
NLP (Text Processing) Generating chatbot responses and text augmentation.
Finance Simulating fraud detection datasets.
The generator and discriminator are designed following Goodfellow et al. [41], ensuring ac-
curate implementation of adversarial training. StyleGAN2 and WGAN-GP architectures are
considered for improved data fidelity. Following Kingma and Welling [42], VAEs encode
data into latent spaces and reconstruct them for generating synthetic samples. The recon-
struction loss and KL divergence are optimized to maintain data distribution integrity.
Step-by-Step Implementation:
• Configure the GAN generator and discriminator architectures using TensorFlow or
PyTorch.
• Train the GAN to generate synthetic data by optimizing the generator and discriminator.
• Use VAEs to encode real data into a compressed latent space and decode them back to
synthesize new data.
• Perform quality validation to ensure that the synthetic data matches the statistical
properties of the original dataset.
Example Code for GAN Training:
import tensorflow as tf
# Define the generator
generator = tf.keras.Sequential([
tf.keras.layers.Dense(128, activation=’relu’),
tf.keras.layers.Dense(256, activation=’relu’),
tf.keras.layers.Dense(784, activation=’sigmoid’)
])
# Define the discriminator
discriminator = tf.keras.Sequential([
tf.keras.layers.Dense(256, activation=’relu’),
tf.keras.layers.Dense(128, activation=’relu’),
tf.keras.layers.Dense(1, activation=’sigmoid’)
])
# Compile the GAN
generator.compile(optimizer=’adam’, loss=’binary_crossentropy’)
discriminator.compile(optimizer=’adam’, loss=’binary_crossentropy’)
sensitivity = 1.0
noise_scale = sensitivity / epsilon
noise = np.random.normal(0, noise_scale, data.shape)
return data + noise
By implementing these steps, the framework ensures robust and practical applicability
across multiple domains.
Privacy Requirement HIPAA (USA) GDPR (EU) CCPA (California) Implemented Technique
§164.502(b)—Minimum Art. 5(1)(c)—Data §1798.100(c)—Data Synthetic Data Generation
Data Minimization
Necessary Rule Minimization Minimization (GANs/VAEs)
§164.312(a)(2)(iv)—Data Art. 32(1)(a)—Encryption §1798.150(a)(1)—Data AES-256 Encryption and
Encryption and Security
Encryption of Data Security TLS 1.3
Recital
§164.514(b)(2)—De- Not Explicitly Required, Differential Privacy
Differential Privacy 26—Anonymization
identification but Recommended Mechanisms
Techniques
Art. 15–22—Right to Privacy-Preserving Query
Data Subject Rights Not Explicitly Required §1798.105—Right to Delete
Access/Erasure Mechanisms
4. Method: add_noise(data)
5. Input: data (Tensor)
6. sensitivity = 1.0
7. noise_scale = sensitivity/self.privacy_budget
8. noise = torch.normal(0, noise_scale, data.shape)
9. return data + noise
The healthcare sector stands out for its robustness, while the financial sector excels in
sensitivity. Analog systems achieving accuracies above 0.96 typically prioritize complexity
over efficiency. While this leads to marginally higher accuracy, it often increases computa-
tional overhead. Our framework balances performance and scalability, making it suitable
for real-time applications. Future iterations will explore advanced architectures and hybrid
models to bridge the accuracy gap while maintaining efficiency.
In the smart city infrastructure, real-time anomaly detection with latency <100 ms was
demonstrated, leveraging edge computing devices optimized for high-speed, low-latency
operations. In contrast, the financial institution scenario achieved a processing time of 1.5 s
per transaction due to higher data complexity and centralized processing. The disparity in
latency is primarily attributed to differences in deployment hardware and data throughput.
The smart city use case involved distributed edge nodes processing smaller data packets
(e.g., IoT sensor data) in parallel. In contrast, the financial scenario relied on centralized
servers analyzing transaction logs with higher computational demands.
Achieving exceptionally high accuracy (>0.99) in anomaly detection often necessitates
more complex models, leading to increased computational costs. This trade-off is particu-
Computers 2025, 14, 55 15 of 18
larly evident in deep learning-based anomaly detection models, which require extensive
GPU processing, higher memory allocation, and prolonged training times. Table 5 com-
pares different model architectures used in this study, highlighting their accuracy–resource
trade-offs.
Model Accuracy Training Time (h) GPU Memory (GB) CPU Usage (%)
Random Forest 0.92 0.5 N/A 45%
SVM 0.94 1.2 N/A 55%
LSTM (Deep Learning) 0.97 3.8 6 GB 70%
GAN-based Framework (Ours) 0.96 2.5 4 GB 65%
High-Complexity CNN 0.99+ 8.2 12 GB 85%
From Table 5, we can observe that achieving an accuracy of >0.99, particularly with
deep CNN-based anomaly detection models, results in a significant increase in GPU
memory usage (12 GB) and training time (8.2 h). Our GAN-based framework maintains a
balance between accuracy (0.96) and computational efficiency, requiring only 4 GB GPU
memory and training within 2.5 h, making it a scalable and practical alternative.
Author Contributions: Conceptualization, G.S.N. and K.M.; methodology, H.G.; software, G.S.S. and
M.M.; validation, G.S.N., K.M. and H.G.; formal analysis, G.S.N. and M.M.; investigation, M.H.M.
and S.R.A.; resources, K.M. and Yadulla, A.R.Y.; data curation, H.G. and M.M.; writing—original
draft preparation, G.S.N. and M.H.M.; writing—review and editing, K.M. and M.M.; visualization,
H.G. and M.H.M.; supervision, G.S.N. and K.M.; project administration, H.G. and G.S.N.; fund-
ing acquisition, M.H.M. and S.R.A. All authors have read and agreed to the published version of
the manuscript.
Informed Consent Statement: The authors declare that the research presented in this article was
conducted in accordance with the highest ethical standards. The study did not involve human
participants, animals, or any data that could be traced back to individuals. All data used in this
research were publicly available or generated through the experimental setups described in the paper.
Data Availability Statement: The source code and additional information used to support the
findings of this study are available from the corresponding author upon request.
Acknowledgments: The authors would like to acknowledge the support and resources provided
by the University of the Cumberlands. The authors are grateful for the conducive environment that
allowed for the successful completion of this study.
References
1. Al-Nemrat, A.; Rahman, M. AI-based cybersecurity solutions for evolving cyber threats. J. Inf. Secur. Appl. 2021, 58, 102786.
[CrossRef]
2. Abomhara, M.; Køien, G.M. Cybersecurity and the internet of things: Vulnerabilities, threats, intruders and attacks. J. Cybersecur.
Priv. 2019, 1, 47–79. [CrossRef]
3. Amine, A.M.; Chakir, E.M.; Issam, T.; Khamlichi, Y.I. A Review of Cybersecurity Management Standards Applied in Higher
Education Institutions. Int. J. Saf. Secur. Eng. 2023, 13, 1109–1116. [CrossRef]
4. Alcaraz, C.; Lopez, J.; Wolthusen, S.D. Security and privacy in distributed edge computing. IEEE Internet Things J. 2020, 7,
9998–10010. [CrossRef]
5. Ali, A.; Hossain, M.A.; Islam, R. Privacy-preserving machine learning: Threats and solutions. IEEE Access 2022, 10, 57652–57666.
[CrossRef]
6. Anderson, R.; Kuhn, M.G. Tamper resistance—A cautionary note. In Proceedings of the 2nd USENIX Workshop on Electronic
Commerce, Oakland, CA, USA, 18–21 November 1996.
7. Biswas, S.; Mollah, M.B. Generative adversarial networks for cybersecurity: A comprehensive survey. IEEE Access 2021, 9, 129073–129087.
[CrossRef]
8. Cao, Y.; Zhu, H.; Wang, Y.; Liang, Z. Federated learning-based cybersecurity framework for data privacy management. IEEE Trans.
Inf. Forensics Secur. 2023, 18, 1234–1246. [CrossRef]
9. Chaudhry, J.A.; Mahmood, H. Data privacy and security challenges in cloud computing. IEEE Access 2020, 8, 116139–116145.
[CrossRef]
10. Chen, H.; Hu, F. AI-enhanced cybersecurity: Threat detection and mitigation using deep learning. IEEE Commun. Surv. Tutor.
2021, 23, 2621–2651. [CrossRef]
11. Chowdhury, M.; Ferdous, M.S.; Biswas, S. Enhancing cybersecurity with blockchain-based decentralized data privacy manage-
ment. J. Netw. Comput. Appl. 2022, 190, 103190. [CrossRef]
12. Golda, A.; Mekonen, K.; Pandey, A.; Singh, A.; Hassija, V.; Chamola, V.; Sikdar, B. Privacy and Security Concerns in Generative
AI: A Comprehensive Survey. IEEE Access 2024, 12, 48126–48144. [CrossRef]
13. Deb, P.; Ghosh, S. AI-driven data privacy and cybersecurity in the cloud environment. IEEE Trans. Cloud Comput. 2021, 9, 808–820.
[CrossRef]
14. Diba, K.; Faghih, F. Generative AI models for cybersecurity: Applications and challenges. IEEE Secur. Priv. 2021, 19, 42–50.
[CrossRef]
15. Fernandes, L.; Rodrigues, J.J.P.C. Edge computing in cybersecurity: Challenges and opportunities. IEEE Access 2020,
8, 21498–21508. [CrossRef]
16. Gao, L.; Zhu, H. Data privacy in AI-based cybersecurity frameworks: A survey. J. Syst. Archit. 2022, 127, 102408. [CrossRef]
Computers 2025, 14, 55 18 of 18
17. Hasan, M.; Islam, M.R. Generative adversarial networks in cybersecurity: A comprehensive review. IEEE Access 2019, 7,
85170–85184. [CrossRef]
18. He, Y.; Xu, L. Blockchain-based privacy-preserving AI for enterprise data security. IEEE Trans. Ind. Informatics 2021, 17, 8194–8203.
[CrossRef]
19. Hussain, F.; Khan, M.A. AI-enhanced cybersecurity for enterprise data privacy: Frameworks and approaches. J. Cybersecur. Priv.
2023, 5, 56–74. [CrossRef]
20. Jain, A.; Gupta, M. Generative AI models for cybersecurity in enterprise environments. ACM Comput. Surv. 2022, 55, 1–28.
[CrossRef]
21. Jiang, Y.; Li, W. AI-based cybersecurity solutions for privacy-preserving enterprise data management. IEEE Trans. Inf. Forensics
Secur. 2020, 15, 3357–3369. [CrossRef]
22. Joo, M.; Kim, H. Machine learning and AI for cybersecurity in the enterprise: Privacy and security implications. IEEE Commun.
Surv. Tutor. 2023, 25, 953–976. [CrossRef]
23. Kammoun, M.; Chelly, H. Cybersecurity and data privacy management using AI: Challenges and solutions. IEEE Access 2022, 10,
57685–57698. [CrossRef]
24. Kang, H.; Park, J. AI-driven cybersecurity solutions for enterprise data protection. IEEE Trans. Cloud Comput. 2020, 8, 1043–1055.
[CrossRef]
25. Khan, M.A.; Hussain, F. Cybersecurity frameworks for AI-enhanced data privacy in enterprises. J. Inf. Secur. Appl. 2023, 68, 103192.
[CrossRef]
26. Shaik, I.; Chandran, N.; A, R.M. Privacy and data protection in the enterprise world. CSIT 2022, 10, 37–45. [CrossRef]
27. Li, J.; Chen, C. AI-powered privacy-preserving cybersecurity solutions for enterprises. J. Netw. Comput. Appl. 2020, 163, 102654.
[CrossRef]
28. Liu, Y.; Zhu, H. AI-enhanced data privacy management in enterprise cybersecurity. IEEE Trans. Inf. Forensics Secur. 2023, 18,
2345–2356. [CrossRef]
29. Luo, X.; Wu, D. AI-based enterprise cybersecurity: Challenges and prospects. J. Syst. Softw. 2022, 191, 110287. [CrossRef]
30. Mahmood, H.; Abdullah, A. Generative AI techniques for enhancing cybersecurity in enterprise networks. IEEE Trans. Ind. Infor.
2021, 17, 4053–4065. [CrossRef]
31. Mollah, M.B.; Biswas, S. Blockchain-based AI frameworks for enterprise data privacy management. IEEE Trans. Eng. Manag. 2021,
69, 864–878. [CrossRef]
32. Nasr, M.; Shokri, R. Privacy-preserving AI in cybersecurity: Techniques and challenges. IEEE Commun. Surv. Tutor. 2022, 24,
3456–3481. [CrossRef]
33. Phan, T.Q.; Do, H. Data privacy and AI-enhanced cybersecurity in cloud computing. IEEE Access 2020, 8, 103910–103920.
[CrossRef]
34. Prasad, S.; Mishra, A. AI-enhanced cybersecurity frameworks for enterprise data privacy. IEEE Trans. Inf. Forensics Secur. 2021, 16,
3357–3371. [CrossRef]
35. Qadir, J.; Ahmad, S. Federated learning for cybersecurity and data privacy in enterprises. IEEE Access 2023, 11, 31245–31260.
[CrossRef]
36. Ren, Y.; Xu, L. Generative adversarial networks for enterprise cybersecurity: A review. IEEE Commun. Surv. Tutor. 2021, 23,
154–176. [CrossRef]
37. Shao, X.; Yang, L. AI-driven cybersecurity strategies for enterprise data protection. J. Inf. Secur. Appl. 2020, 52, 102489. [CrossRef]
38. Smith, A.; Jones, B. Data privacy challenges in AI-based cybersecurity frameworks. IEEE Access 2021, 9, 154573–154590. [CrossRef]
39. Wang, W.; Liu, Y. AI and blockchain for cybersecurity in enterprise data privacy management. IEEE Trans. Cloud Comput. 2022, 10,
154–168. [CrossRef]
40. Xu, Y.; Han, J. AI-enhanced cybersecurity frameworks for data privacy in enterprise environments. IEEE Trans. Eng. Manag. 2023,
69, 1456–1468. [CrossRef]
41. Zhang, C.; Wu, X. A survey of AI-enhanced cybersecurity techniques for data privacy management. IEEE Commun. Surv. Tutor.
2023, 25, 1769–1795. [CrossRef]
42. Koza, E. Semantic analysis of ISO/IEC 27000 standard series and NIST cybersecurity framework to outline differences and
consistencies in the context of operational and strategic information security. Med. Eng. Themes 2022, 2, 26–39.
43. Mirsky, Y.; Lee, W. The creation and detection of deepfakes: A survey. ACM Comput. Surv. (CSUR) 2021, 54, 1–41. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.