Securing AI · · 7 min read

Securing Generative AI Systems

This essay explores the unique security considerations surrounding generative AI and why I believe this will be the next significant frontier in cybersecurity.

Securing Generative AI Systems
GenAISec by Philip Dursey and leonardo.ai, the AI Security Pro human machine (rendering) team

As a founder and CEO of an AI security company, with recent experience as an AI CISO, I've observed the rapid evolution of generative AI with both excitement and concern. The potential of these systems is immense, but so are the security challenges they present. This essay explores the unique security considerations surrounding generative AI and why I believe this will be the next significant frontier in cybersecurity.

The Unique Nature of Generative AI

Generative AI systems, such as large language models (LLMs), differ fundamentally from traditional software. Unlike conventional software that follows explicit, human-written rules, generative AI learns patterns from vast datasets and generates outputs based on statistical likelihoods [1]. This creates a new paradigm for security, where traditional vulnerabilities like buffer overflows or SQL injections may not apply, and new, unforeseen risks emerge.

The Challenge of Determinism

A core principle in cybersecurity is determinism—ensuring systems behave predictably to secure them effectively. Generative AI introduces non-determinism, meaning the same input can yield different outputs. This complicates traditional security testing methodologies [2]. For example, creating unit tests for generative AI is challenging because there is no single "correct" output to compare against. We must develop new testing methodologies to ensure these systems are secure.

Provable Security and Its Limits

Provable security involves mathematically verifying that a system adheres to specific security properties. This approach, while rigorous, faces challenges when applied to generative AI due to the inherent complexity and variability of these systems. Unlike traditional software, where formal methods can prove the absence of certain vulnerabilities, the dynamic and evolving nature of AI models makes it difficult to establish comprehensive proofs of security. As AI models grow in complexity, the gap between what can be proven and the actual behavior of the system widens, leaving room for potential vulnerabilities [8].

The Inadequacy of Defense-in-Depth

The traditional approach to cybersecurity often relies on defense-in-depth, a strategy that uses multiple layers of defense to protect systems. However, this approach may be inadequate for generative AI systems. The dynamic nature of these models, coupled with the vast attack surface they present, can overwhelm traditional defensive measures. For example, while firewalls and intrusion detection systems are effective against known threats, they may not detect subtle manipulations or emergent behaviors in AI outputs that could signify an attack.

Exploring Alternatives: Threat Perception Management

Given the limitations of traditional security paradigms, alternative approaches such as threat perception management are gaining traction. Threat perception management focuses on understanding and influencing the perceptions and behaviors of potential attackers. This approach involves:

  • Adaptive Threat Models: Continuously updating threat models based on evolving tactics, techniques, and procedures (TTPs) used by attackers. This ensures that security measures remain relevant in the face of changing threats [10].
  • Behavioral Analytics: Leveraging AI and machine learning to analyze patterns in data, identifying anomalous behavior that may indicate a security threat. This method is particularly suited to dealing with the non-deterministic nature of generative AI systems, as it can adapt to new and unexpected attack vectors.
  • Adaptive Deception Technologies: Implementing deceptive practices, such as honeypots and adaptive decoy systems, to mislead attackers and study their tactics without compromising critical systems. These tools can provide valuable insights into attacker behavior and help refine threat models.
  • Dynamic Risk Assessment: Moving away from static security postures to a more dynamic assessment of risk, continuously adjusting defenses based on real-time data and emerging threats. This approach aligns with the fast-paced nature of generative AI development and deployment.

These alternatives emphasize a proactive and adaptive security posture, essential for addressing the unique challenges posed by generative AI systems. As the field of AI security evolves, integrating these approaches with traditional cybersecurity measures can provide a more comprehensive and resilient defense strategy.

The Data Dilemma

The quality and security of generative AI systems are highly dependent on the data they are trained on. This presents two primary challenges:

  • Data Poisoning: Attackers could introduce malicious data into training sets, potentially influencing model outputs. This is especially concerning in critical infrastructure contexts, where subtle manipulations could introduce vulnerabilities in generated code or configurations.
  • Data Privacy: These models can inadvertently memorize and reproduce sensitive information from their training data, raising significant privacy concerns. For instance, researchers demonstrated in 2021 that GPT-2 could output individual training examples, including personally identifiable information [3]. This underscores the need for robust data sanitization and privacy-preserving training techniques.

Synthetic Data Generation for Attack Surface Manipulation

One proactive measure to address the data dilemma is the use of synthetic data generation. Synthetic data can be used to augment training datasets, improving model robustness and diversity. More importantly, it can be strategically employed to manipulate the attack surface, confusing potential attackers about the true nature of the training data [11].

For example, by including carefully crafted synthetic data points designed to mislead adversaries, organizations can obscure the real patterns within the data. This approach complicates the task for attackers attempting to poison data or extract sensitive information, as they cannot easily discern real from synthetic data.

Adversarial AI Data Poisoning as an Active Defense

Adversarial AI techniques, often discussed in the context of attacking AI systems, can also be repurposed for defense. One such strategy is to intentionally poison the data used by attackers. This concept, known as adversarial AI data poisoning, involves inserting "trap" data into public datasets or decoy systems that attackers might use to train their models. 

When attackers unknowingly use this poisoned data, their AI systems learn incorrect patterns or biases, reducing their effectiveness in launching attacks. This strategy turns the tables, using the attackers' reliance on data against them. For example, if an attacker trains a model on data that has been subtly altered to produce incorrect outputs under certain conditions, the resulting model might fail to operate correctly when deployed, neutralizing its threat [12].

Data Protection

Addressing the data dilemma in generative AI involves a multifaceted approach, combining rigorous data validation, advanced privacy-preserving techniques, and innovative defensive strategies like synthetic data generation and adversarial AI data poisoning. By proactively managing the quality and integrity of training data, we can significantly enhance the security and reliability of generative AI systems. As these systems become more integral to various industries, ensuring the security of their data pipelines will be crucial in safeguarding against both traditional and emerging threats [13].

The Inference Attack Vector

Generative AI introduces new attack vectors during the inference process. Adversarial attacks, previously studied in image classification, are now relevant for generative models. Attackers could craft inputs to cause models to generate harmful or unintended outputs, ranging from malicious code to misinformation [4]. For example, "universal adversarial triggers" can manipulate NLP models to produce specific outputs regardless of the input context, posing significant security risks. We'll continue to explore this theme in subsequent articles, beyond what we've described in Navigating the Complex Landscape of AI Security: Threats and Countermeasures in Artificial Intelligence Systems.

The Scale Problem

The sheer scale of generative AI systems, often involving billions of parameters and trained on terabytes of data, presents unique security challenges [5]. Traditional security approaches, such as code review, are impractical at this scale. For instance, GPT-3 has 175 billion parameters, making comprehensive security analysis daunting. The centralization of AI model development among a few organizations also creates a single point of failure, increasing the potential impact of high-impact, cascading, security vulnerabilities.

The Ethical Dimension

Security in generative AI extends beyond technical concerns to ethical issues. These systems can perpetuate and amplify societal biases present in training data, leading to biased or harmful outputs [6]. Ensuring ethical behavior in AI systems is crucial, as unethical outputs can lead to real-world consequences, such as discrimination or the spread of misinformation.

The Way Forward

To secure generative AI systems, we need to focus on several key areas:

  • Robust Training Techniques: Developing training methodologies resistant to data poisoning and that produce more stable and predictable models. Techniques like differential privacy and federated learning are promising approaches.
  • Advanced Testing Frameworks: Creating new testing frameworks to identify vulnerabilities in generative AI systems. OpenAI and Google's AI Red Teams are examples of proactively addressing machine learning risks.
  • Secure Deployment Architectures: Implementing secure deployment practices, including secure enclaves and differential privacy, to protect models and data. OpenAI's secure APIs for accessing language models exemplify this approach.
  • Ethical Guidelines and Governance: Establishing clear guidelines and governance structures for AI development and deployment, considering security and ethical implications. Initiatives like the EU's proposed AI Act are steps in the right direction.
  • Interdisciplinary Collaboration: Encouraging collaboration between AI researchers, cybersecurity experts, ethicists, and policymakers to address the multifaceted challenges of AI security [7].

Conclusion

Securing generative AI systems is a complex challenge that extends beyond technical measures to include ethical and societal considerations. As these systems become more integrated into our digital infrastructure, their security becomes increasingly critical. The rise of cloud computing created a new field of cloud security, and similarly, generative AI will drive the development of new security tools, methodologies, and companies. For those in cybersecurity, this is a call to action to address these challenges proactively, ensuring a secure and ethical AI-driven future.


References:

1. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). "Language models are unsupervised multitask learners." *OpenAI Blog*, 1(8), 9.

2. Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). "On the dangers of stochastic parrots: Can language models be too big?" *Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency*, 610-623.

3. Carlini, N., Tramer, F., Wallace, E., Jagielski, M., Herbert-Voss, A., Lee, K., ... & Raffel, C. (2021). "Extracting training data from large language models." *30th USENIX Security Symposium (USENIX Security 21)*, 2633-2650.

4. Wallace, E., Feng, S., Kandpal, N., Gardner, M., & Singh, S. (2019). "Universal adversarial triggers for attacking and analyzing NLP." *arXiv preprint arXiv:1908.07125*.

5. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., ... & Amodei, D. (2020). "Language models are few-shot learners." *arXiv preprint arXiv:2005.14165*.

6. Weidinger, L., Mellor, J., Rauh, M., Griffin, C., Uesato, J., Huang, P. S., ... & Gabriel, I. (2021). "Ethical and social risks of harm from Language Models." *arXiv preprint arXiv:2112.04359*.

7. Brundage, M., Avin, S., Clark, J., Toner, H., Eckersley, P., Garfinkel, B., ... & Amodei, D. (2018). "The malicious use of artificial intelligence: Forecasting, prevention, and mitigation." *arXiv preprint arXiv:1802.07228*.

8. Katz, G., Barrett, C., Dill, D. L., Julian, K., & Kochenderfer, M. J. (2017). "Reluplex: An efficient SMT solver for verifying deep neural networks." *arXiv preprint arXiv:1702.01135*.

9. Osman, H., & Nowozin, S. (2018). "Variational Bayesian dropout: Pitfalls and fixes." *arXiv preprint arXiv:1808.02228*.

10. Biggio, B., Nelson, B., & Laskov, P. (2012). "Poisoning attacks against support vector machines." *29th International Conference on Machine Learning (ICML)*.

11. Antoniou, A., Storkey, A., & Edwards, H. (2017). "Data augmentation generative adversarial networks." *arXiv preprint arXiv:1711.04340*.

12. Kwon, H., Kim, M., Yoon, S., & Hwang, S. J. (2021). "Adversarial Poisoning Attacks on Neural Networks via Meta Learning." *Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security (CCS '21)*.

13. Shokri, R., & Shmatikov, V. (2015). "Privacy-preserving machine learning as a service." *Proceedings of the 2015 ACM SIGSAC Conference on Computer and Communications Security (CCS '15)*, 1310-1321.

Read next