Securing AI · · 8 min read

AI Supply Chain Security

Recent incidents have highlighted the extensive impact of supply chain attacks, affecting not only traditional software systems but also AI-specific contexts.

AI Supply Chain Security
AI Supply Chain Attack by Philip Dursey and leonardo.ai, the AI Security Pro human machine (rendering) team

Introduction

With the accelerated evolution of artificial intelligence (AI), the importance of securing the AI supply chain is increasingly evident. The complex ecosystem supporting the development, deployment, and operation of AI systems presents an expansive attack surface for malicious actors. Recent incidents have highlighted the extensive impact of supply chain attacks, affecting not only traditional software systems but also AI-specific contexts.

The SolarWinds hack of 2020 and the Kaseya ransomware attack in 2021 set the stage for a new era of supply chain vulnerabilities. While these incidents did not specifically target AI systems, they illustrate potential risks within complex technological supply chains—risks amplified in the AI domain due to the unique characteristics and requirements of intelligent systems.

This essay explores the key elements of AI supply chain security, focusing on hardware vulnerabilities, boot security, and AI-specific attack vectors. We will examine the associated risks, discuss mitigation strategies, and consider the impact of emerging technologies and regulations on the security landscape.

The AI Supply Chain: An Expanded Overview

  1. Data Acquisition and Preparation
  2. Model Development and Training
  3. Software Dependencies
  4. Hardware and Infrastructure
  5. Deployment and Maintenance
  6. Continuous Learning and Updates
  7. Third-party Model Integration: The integration of pre-trained models or components from external sources introduces new supply chain risks.
  8. Edge AI Deployment: Securing AI models deployed on diverse edge devices has become a significant challenge with the proliferation of edge computing.

Hardware Vulnerabilities in AI Systems

AI systems rely on hardware components, from specialized AI accelerators to traditional CPUs and memory systems, which collectively present a complex attack surface.

  • Side-Channel Attacks: These attacks exploit physical characteristics of hardware to extract sensitive information. Cache-based attacks, such as those exploiting speculative execution vulnerabilities like Spectre and Meltdown, remain significant threats, potentially allowing extraction of AI model parameters.
  • Hardware Trojans and Supply Chain Attacks: The complexity of AI accelerators increases the threat of hardware trojans. To mitigate these risks, advanced supply chain verification techniques are being developed, including "split manufacturing" and "logic locking," which complicate tampering or reverse-engineering efforts.
  • Memory-based Vulnerabilities: The massive memory requirements of modern AI models exacerbate traditional vulnerabilities like row hammer attacks, which could alter AI model weights or input data, causing targeted misclassifications.
  • Quantum Computing: Threat and Opportunity: Although full-scale quantum computers capable of breaking current cryptographic systems are still years away, their potential threat to AI systems is taken seriously. Post-quantum cryptography (PQC) is being integrated into AI infrastructure . Conversely, quantum technologies offer new security possibilities, such as quantum key distribution (QKD) and quantum machine learning algorithms for robust encryption schemes.

Boot Security: The First Line of Defense

Ensuring boot security is essential for maintaining overall system integrity, particularly in AI systems handling sensitive data or critical functions.

  • Advanced Persistent BIOS Threats (APBT): Sophisticated UEFI rootkits have been discovered that can persist across OS reinstalls, potentially targeting AI development environments .
  • Secure Boot Enhancements: Secure Boot implementations have been strengthened, extending the chain of trust to firmware, bootloaders, and initial ramdisk components, offering comprehensive protection against boot-time attacks.
  • Remote Attestation for Distributed AI: With the rise of federated learning and distributed AI systems, remote attestation has become crucial. New protocols provide standardized ways for distributed AI nodes to verify their integrity before engaging in collaborative tasks.

AI-Specific Attack Vectors and Mitigations

  • Model Theft and Intellectual Property Protection: As AI models grow in value, attacks aimed at stealing or reverse-engineering them have become more sophisticated. Techniques such as watermarking, encryption-in-use, and federated learning are being developed to protect models and intellectual property.
  • Adversarial Attacks in Deployment: These attacks can cause unexpected and potentially dangerous behaviors in AI systems. Mitigation strategies include adversarial training, runtime monitoring, and ensemble methods to reduce the impact of specific model vulnerabilities .

Regulatory Landscape and Compliance

The regulatory environment for AI has evolved significantly. The EU's AI Act includes specific requirements for AI supply chain security, while in the US, NIST provides guidelines for securing AI systems throughout their lifecycle. Key regulatory requirements include:

1. Mandatory security audits for high-risk AI systems

2. Supply chain transparency and traceability

3. Incident reporting for AI security breaches

4. Privacy-preserving techniques for AI training and inference

Cloud AI Security

As cloud platforms become the primary environment for many AI workloads, securing these environments is crucial to protect sensitive data, proprietary models, and the overall integrity of AI systems. Cloud AI security encompasses several advanced measures tailored to the unique challenges posed by the cloud infrastructure:

  1. Confidential Computing: This technology provides hardware-based secure enclaves, which isolate data and computations from the rest of the system. Even the cloud service provider cannot access the data or computations occurring within these enclaves. This is particularly important for AI workloads that handle sensitive data, such as healthcare or financial information. By using confidential computing, organizations can ensure data privacy and compliance with regulations even when using third-party cloud services.
  2. AI-specific Identity and Access Management (IAM): Traditional IAM systems are being adapted to meet the needs of AI workflows. These specialized IAM systems offer granular control over who can access AI models, data, and computational resources. They allow organizations to define specific roles and permissions for different users, ensuring that only authorized personnel can modify or use AI systems. This is crucial for protecting against unauthorized access, data leaks, and potential misuse of AI capabilities.
  3. Secure Multi-party Computation (SMPC): SMPC enables multiple parties to collaboratively perform computations on their data without revealing the actual data to one another. This is especially valuable in scenarios where AI models are trained or inferred using datasets from different organizations or jurisdictions. By ensuring data remains confidential throughout the computation process, SMPC facilitates secure collaboration and model development across entities with varying data privacy requirements, as highlighted recently here: The Hidden Battlefield: Model Hijacking in the Age of Collaborative AI
  4. Data Encryption and Secure Storage: Data encryption in transit and at rest is a foundational aspect of cloud security. For AI workloads, this includes not only encrypting data but also model parameters, metadata and configurations. Secure storage solutions, such as encrypted databases and storage services, are employed to protect sensitive information from unauthorized access or breaches. Additionally, key management systems ensure that encryption keys are securely stored and managed, reducing the risk of key compromise.
  5. Network Security and Microsegmentation: Cloud environments offer advanced network security features, such as virtual private clouds (VPCs), cross-cloud VPCs, firewalls, and microsegmentation. These features help isolate different parts of an AI system, reducing the risk of lateral movement by attackers within the cloud infrastructure. Microsegmentation, in particular, allows for the creation of secure zones within a cloud environment, ensuring that even if one segment is compromised, the others remain secure.
  6. Monitoring and Incident Response: Cloud providers offer comprehensive monitoring tools that allow organizations to track the performance and security of their AI systems in real-time. These tools can detect unusual activities, such as unauthorized access attempts or abnormal data usage patterns, and trigger alerts for further investigation. Additionally, cloud providers often offer incident response services to help organizations respond quickly and effectively to security incidents. There is an emerging class of cloud-centric ai-driven adaptive defense solutions to support rapid response and remediation.
  7. Compliance and Auditing: Cloud AI environments must comply with various regulatory frameworks and industry standards, such as GDPR, HIPAA, and ISO/IEC 27001. Cloud providers typically offer compliance certifications and audit trails that help organizations demonstrate adherence to these regulations. These features are critical for organizations operating in highly regulated sectors, ensuring that their AI operations are compliant with legal and industry requirements.

Bare-Metal HPC (High-Performance Compute) Security Advantages

Bare-metal high-performance computing (HPC) environments offer distinct security advantages over virtualized cloud environments, especially for organizations handling sensitive AI workloads. Bare-metal servers provide dedicated hardware resources, eliminating the hypervisor layer, which is a potential attack vector in virtualized environments. Here are some key security benefits of using bare-metal HPC for AI, similar to what we've build at Hydra Host:

  1. Isolation and Performance: Bare-metal servers provide complete isolation from other tenants, which is critical for security-sensitive applications. This isolation reduces the risk of data leakage and side-channel attacks that can occur in multi-tenant cloud environments. Additionally, the absence of virtualization overhead allows AI workloads to run with maximum performance, which is crucial for compute-intensive tasks such as training large AI models.
  2. Enhanced Control and Customization: Organizations have full control over the hardware and software stack in bare-metal environments. This control allows for the implementation of customized security measures, such as tailored BIOS configurations, specific operating system hardening practices, and the deployment of proprietary security tools. This level of customization can be essential for meeting stringent security and compliance requirements.
  3. Reduced Attack Surface: By eliminating the hypervisor, bare-metal servers have a reduced attack surface compared to virtualized environments. Hypervisors can be complex and, if compromised, can potentially provide attackers with access to all virtual machines running on the hardware. The simplicity of bare-metal environments thus offers fewer opportunities for attackers to exploit vulnerabilities.
  4. Physical Security: Bare-metal servers are often housed in secure data centers with stringent physical security measures, including controlled access, surveillance, and disaster recovery plans. This physical security complements cybersecurity measures, providing a robust overall security posture for sensitive AI workloads.
  5. Compliance and Data Sovereignty: For organizations in regulated industries or those dealing with sensitive data, bare-metal HPC environments can provide better compliance with data sovereignty laws and regulations. Organizations can choose data center locations to ensure data remains within specific geographic and legal boundaries, complying with local data protection laws and regulations.

By leveraging the advantages of bare-metal HPC environments, organizations can enhance the security of their AI workloads, ensuring that sensitive data and intellectual property are protected against advanced threats. This approach is particularly beneficial for industries such as finance, healthcare, and government, where data security and regulatory compliance are paramount.

Mitigation Strategies: A Holistic Approach

Securing the AI supply chain requires a comprehensive strategy:

1. Implement "Security by Design" in AI Development

2. Enhance Data Security and Governance

3. Secure the Software Supply Chain

4. Harden Hardware Infrastructure

5. Enhance Boot and Runtime Security

6. Protect Models and Intellectual Property

7. Prepare for Post-Quantum Threats

8. Foster a Security-Aware Culture

Conclusion and Future Outlook

As AI systems are increasingly integrated into critical infrastructure and decision-making processes, securing the AI supply chain is crucial for national and economic security. The challenges are complex and evolving, requiring constant vigilance and innovation.

Several trends are poised to influence the future of AI supply chain security:

  1. Quantum-Resistant AI: Development of AI systems inherently resistant to quantum computing threats.
  2. AI-Powered Security: Sophisticated AI systems dedicated to detecting and mitigating security threats in real time, as with HypergameAI's A-TIER platform.
  3. Decentralized AI: Growth of federated and edge AI systems, distributing both computational load and security risks.
  4. Regulatory Harmonization: Efforts to create global standards for AI security, facilitating international collaboration and trade.
  5. Ethical AI Security: Increasing focus on ensuring that security measures are ethically implemented and do not introduce new biases or vulnerabilities.

By adopting a holistic approach to AI supply chain security, leveraging emerging technologies, and fostering collaboration between industry, academia, and government, we can build a more secure foundation for the transformative potential of AI. Ensuring the integrity, confidentiality, and availability of these systems at every step of their lifecycle will be crucial to realizing the full benefits of this revolutionary technology.


References:

1. Jibilian, I., & Canales, K. (2021). The US is readying sanctions against Russia over the SolarWinds cyber attack. Here's a simple explanation of how the massive hack happened and why it's such a big deal. Business Insider.

2. Collier, K. (2021). Kaseya ransomware attack sets off race to hack service providers -researchers. Reuters.

3. Lipp, M., et al. (2018). Meltdown: Reading kernel memory from user space. 27th USENIX Security Symposium.

4. Kocher, P., et al. (2019). Spectre attacks: Exploiting speculative execution. 2019 IEEE Symposium on Security and Privacy (SP).

5. Kwong, A., et al. (2020). RAMBleed: Reading Bits in Memory Without Accessing Them. IEEE S&P 2020.

6. NIST. (2022). Post-Quantum Cryptography Standardization. National Institute of Standards and Technology.

7. Pirandola, S., et al. (2020). Advances in quantum cryptography. Advances in Optics and Photonics, 12(4), 1012-1236.

8. ESET Research. (2018). LoJax: First UEFI rootkit found in the wild, courtesy of the Sednit group. WeLiveSecurity.

9. Tramèr, F., et al. (2016). Stealing Machine Learning Models via Prediction APIs. USENIX Security Symposium.

10. Carlini, N., & Wagner, D. (2017). Towards evaluating the robustness of neural networks. 2017 IEEE Symposium on Security and Privacy (SP).

11. European Commission. (2021). Proposal for a Regulation laying down harmonised rules on artificial intelligence. EUR-Lex.

12. NIST. (2021). AI Risk Management Framework: Initial Draft. National Institute of Standards and Technology.

13. Google Cloud. (2021). Confidential Computing. Google Cloud website.

14. Bonawitz, K., et al. (2019). Towards federated learning at scale: System design. Proceedings of Machine Learning and Systems, 1, 374-388.

15. Papernot, N., et al. (2017). Practical black-box attacks against machine learning. Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security.

Read next