Introduction
Artificial intelligence (AI) and machine learning (ML) systems allow organisations to improve their efficiency in many areas. These systems can help inform decisions, streamline processes and improve customer experience.
Adopting AI and ML systems introduces unique supply chain risks, which can threaten the cyber security of an organisation if not securely managed. Using pre-trained models and third-party datasets is beneficial, but it can also bring existing compromise and supply chain risks. Organisations should know what to look out for when developing or incorporating AI and ML into their systems. Wider AI development advice can be found in the UK National Cyber Security Centre’s Guidelines for secure AI system development.
This guidance is intended for organisations and staff that deploy or develop AI and ML systems and components. This could range from entirely outsourcing an AI system where an organisation only provides the training data, to in-house AI development. When sourcing third-party AI and ML systems and components, the risks and mitigations in this guidance should inform your questions and requirements for the vendor. It aims to:
- highlight the importance of AI and ML supply chain security
- address key risks and mitigations that should be considered when developing or procuring an AI system.
While this guidance provides some mitigation descriptions, organisations should determine which measures to prioritise and how to apply them based on their own risk assessment. Additionally, cyber security staff should still consult detailed technical guidance and frameworks for implementation. To help with this, risks outlined in this guidance are mapped to the National Institute of Standards and Technology’s (NIST) Adversarial Machine Learning (AML) taxonomy where applicable. This guidance in general maps to MITRE’s Adversarial Threat Landscape for Artificial-Intelligence Systems (ATLAS) framework under AI Supply Chain Compromise. Additionally, refer to the More information section at the end of this guidance.
This guidance focuses on aspects unique to AI and ML, though general cyber supply chain risk management still applies. For more information, refer to Cyber supply chain risk management. For an explanation of AI-related terminology in this guidance, refer to Convoluted layers: An artificial intelligence primer. For more information on using AI in general, refer to Engaging with artificial intelligence.
This guidance is being released in collaboration with the following co-sealing agencies:
- Canadian Centre for Cyber Security (Cyber Centre)
- Singapore Cyber Security Agency (CSA)
- Republic of Korea National Intelligence Service (NIS)
- Japan National Cybersecurity Office (NCO)
- New Zealand National Cyber Security Centre (NCSC-NZ)
- United Kingdom National Cyber Security Centre (NCSC-UK)
- United States National Security Agency (NSA)
Overview
The AI and ML supply chain is complex. Vendors and service providers often need to source or manage various components. This includes the underlying AI training data, model, software, infrastructure, hardware, and third-party services.
Each component may contain vulnerabilities and risks that a malicious actor could exploit to compromise confidentiality, availability and integrity of the system. Depending on the component, these exploits could expose training data, degrade AI or ML system functionality, or execute malicious code. This guidance addresses AI and ML broadly, but individual technologies such as LLMs or AI agents may have unique security concerns. Always investigate specific security advice for chosen technologies such as OWASP’s Top 10 for LLM Applications.
Supply chain risk management concepts for AI and machine learning
Organisations should assess their AI and ML supply chain throughout an AI product or service’s lifecycle as part of their broader cyber security risk management strategy. Existing cyber security considerations that would apply to any new software, system or dataset still apply to AI and ML components. For more information on general cyber security management, refer to the Information security manual and Secure by Design practices.
Supply chain integrity assessment
AI and ML considerations should be assessed as part of comprehensive supply chain assessment. For general advice, refer to Identifying cyber supply chain risks. A similar approach is outlined in the Canadian Centre for Cyber Security’s Cyber supply chain: An approach to assessing risk - ITSAP.10.070. Many key considerations are relevant to AI and ML systems such as:
- sensitivity – particularly due to high volumes of data and often wide reach of AI within systems
- ownership – control and influence can affect performance or bias and may change over time
- technical – managing technical threats and understanding new tactics, techniques and procedures (TTPs) that come with AI and ML systems.
Integrating AI and machine learning products and services
Effective risk management requires full visibility of AI and ML systems and their supply chain. Organisations should:
- identify suppliers, manufacturers, distributors, retailers and subcontractors
- ensure systems follow security best practices to reduce impact of future AI or ML integration
- seek information on security controls and policy for AI and ML integrations and consider how it affects their cyber security risk management
- consider how new functionality from AI or ML products could affect their cyber security risk management
- require and maintain an AI Bill of Materials (AIBOM) and Software Bill of Materials (SBOM) or equivalent documentation for AI and ML.
Revise risk management
- Organisations will need to review and revise their risk management to include new risks and controls brought by AI and ML systems. To identify and mitigate threats, vulnerabilities and risks, organisations should:
- conduct initial and ongoing assessments to identify new security situations
- perform threat modelling to reveal potential new attack vectors
- conduct vulnerability mapping and ensure mitigations are enacted
- develop and maintain an incident response plan for your AI supply chain.
Working with third parties
When working with AI and ML vendors and suppliers, discuss cyber security considerations with them at an early stage, and understand shared responsibilities. Organisations should:
When working with AI and ML vendors and suppliers, discuss cyber security considerations with them at an early stage, and understand shared responsibilities. Organisations should:
- execute due diligence on vendors and evaluate their security posture, data handling practices and extended supply chain risk
- ensure direct and third-party suppliers report and respond to vulnerabilities and incidents upon discovery
- prioritise supply chain components or vendors that are responsible for critical cyber security functions
- implement supplier software verification processes
- consider cyber security management for implementation and throughout the product’s life cycle
- include cyber security requirements in contractual agreements and outline shared responsibility models.
Case study: Data leaks from third parties and AI
In 2025 alone, there have been several high profile data leakages due to compromise of third-party services. A recent survey found that the majority of surveyed organisations had been impacted by AI-related data leaks. Another survey reported that several organisations have also faced direct breaches to their AI systems.
AI systems commonly collect and retain large amounts of data for model training or context. This aggregation along with the often personal or sensitive nature of the data makes these systems attractive targets for malicious actors with significant potential impact when breaches occur.
Securing your organisation’s data
Organisations should understand how AI and ML systems affect their data security by:
- considering the value of their data and what other information could be derived from it
- understanding what data they process, store and communicate to support AI or ML systems
- considering who has access to sensitive data, model inventory, data storage, infrastructure or hardware
- confirming and considering which AI and ML vendors can access their data and how they use it
- being aware of data that may be transmitted, processed or stored in a foreign location.
Communicating and training within your organisation
Incorporating AI and ML systems will likely require additional internal management for supply chain and staff training. Organisations should:
- establish clear communication channels for any AI and ML supply chain concerns
- educate all staff involved in AI or ML system development, deployment, use and maintenance about cyber security risks and best practices.
AI data
AI and ML systems typically use large datasets and knowledge bases that often need to be maintained for future retraining or new models. This data may be sensitive and is often stored in a single location so it can be efficiently retrieved when training the model. Organisations may opt to collect their own training data or acquire external datasets – each option coming with varying considerations regarding accuracy and security.
Conventional data security management should still be applied, including consideration of increased data aggregation. For more information, refer to AI data security.
Risks
This section covers some risks associated with data in the AI and ML supply chain.
Low-quality or biased data
Key question: Where do you source your AI training data? How do you ensure it is high quality and unbiased?
NIST AML Mappings: None (not an intentional attack)
Low-quality or biased data may be inaccurate or may not clearly represent the problem the AI or ML system aims to solve. Poor data collection or generation processes can lead to bias or inaccurate labelling, causing the resultant model to be harder to train and less effective overall. This can cause the resulting application to have more incorrect classifications or predictions, ultimately reducing robustness.
Data poisoning
Key question: How do you ensure your training data has not been maliciously tampered with?
NIST AML Mappings: NISTAML.012, .013, .021, .023, .024
Data poisoning involves malicious modifications to data that can affect the final AI or ML model and its outputs or cause unintended system behaviour. Data poisoning can be hard to detect due to the size and complexity of datasets, especially if data is already compromised at the source.
The effects of poisoned data on the resulting model can include:
- degraded performance and bias
- unintended or malicious responses to a specific data feature.
In other cases, the poisoned data can be used for indirect prompt injection. This is where the model mistakes part of the data as an instruction, which can lead to malicious behaviour.
Case study: Organisations affected by data poisoning
Joint reporting on AI data security confirmed that adversarial actors are already poisoning AI training data. Other research from 2025 found that several organisations had been a victim of AI data poisoning.
Training data exposure
Key question: How do you reduce the potential damage from exposed training data?
NIST AML Mappings: NISTAML.031, .032, .033, .034, .037, .038
Through manipulation or reverse engineering of AI and ML models, adversaries may be able to extract content or insights from data used to train the model. Even if the model can no longer access the original data, this information can be derived from the model’s training memory. Attacks of this type include:
- model inversion (causing general insight into data)
- membership inference (if a particular sample is in the data)
- training data extraction.
These attacks could reveal sensitive information or otherwise be a privacy and confidentiality concern. While this is a risk primarily when using the model, organisations can apply several mitigations prior to model deployment to reduce risk.
Mitigations
Preventative measures are often more effective than detection for data risks. Mitigations should focus on ensuring the security, integrity and accuracy of the data from the outset.
Data collection and generation methods
Use a standardised and documented methodology for collection or generation of data. This can help with consistency and reduce the risk of bias, inaccuracy or other issues. Preference or weight “real” data over generated data to avoid bias that synthetic generation processes can cause.
Quarantine and testing
Organisations should quarantine and test externally sourced data before incorporating it into their internal systems. Store the data on dedicated isolated systems to avoid damage from potential malicious content. Use automated testing to scan for issues, such as malicious content, poisoning and watermarks. Only migrate the data to internal systems and repositories after verifying a base level of security.
Data review and preprocessing
Organisations should review and preprocess data before using it in an AI or ML model. Data review should check that data is as expected and free of obvious errors. Preprocessing can help prevent bias by correcting issues such as missing features, duplicates or uneven sample distribution. These steps can help to ensure the data is of sufficient quality, though organisations should take care not to introduce bias themselves.
Data sanitisation
Sanitisation is a data-refining process that can include techniques such as filtering, sampling and normalisation. This process can reduce unwanted aspects in a dataset like noise, outliers and poisoned content.
Trusted sources
Source data from trusted and reputable providers where possible. Avoiding unknown sources can be a simple but effective way to reduce the risk of poisoned or maliciously modified data.
Data verification and provenance
Verification methods (such as checksums, hashes and digital signatures) are a way to check the integrity of digital data. Note that this only verifies integrity against the source and does not include any modifications made to the data prior to publishing.
More robust verification methods (such as lineage tracking and data provenance) are in development that will allow visibility of all changes on a dataset since its creation. With additional chain-of-custody information and tracking of changes, maliciously modified or poisoned data can be identified early. For an example of this in practice, refer to Content credentials: Strengthening multimedia integrity in the generative AI era.
Ensemble methods
Ensemble methods involve training multiple different datasets (or subsets) and models on the same problem to compare performance and consistency. This means that if a dataset or model is compromised, combinations using that component may have significantly different outputs to the others. These metrics can then prompt developers to further investigate, rectify or remove those components, which reduces risks from data or models that are poisoned or of low-quality. Orchestration tools can help automate this process, but as ensemble methods are relatively complex and need more resources, they may not be practical for lower-risk situations.
Training data obfuscation
Obfuscating training data removes identifiable or sensitive features while maintaining key aspects and patterns. Organisations should remove unnecessary or unimportant data and use anonymisation or differential privacy methods where possible. Where sensitive content is required, generating synthetic data based on the real content can reduce sensitivity while maintaining important features. While these methods do not prevent data extraction attacks, the obfuscation can reduce further damage.
Machine learning models
An ML model is the structure that transforms inputs into outputs based on patterns learned from training data. Attacks often involve manipulating the parameters within a trained model to cause an effect such as bias, degraded accuracy, or embedding of malicious code.
Risks
The following are some risks associated with the supply chain of ML models.
Model serialisation and similar attacks
Key question: How do you verify that a saved model is not compromised?
NIST AML Mappings: Part of NISTAML.05
Serialising, archiving and similar methods package ML models and associated components into portable and widely compatible formats. A model serialisation attack could occur when malicious code is added to the model package during the serialisation process. Other methods could add malicious code to a post-install script or preprocessing method, yielding similar effects with minimal user input. Malicious actors can then post the compromised file online to spread malware that activates when a user loads the contents. Malicious code added to associated components can cause similar effects.
Case study: Malicious ML models on sharing platform
Research from 2025 has shown that serialised model files containing malicious content continue to be uploaded to popular sharing platforms without being flagged by security checks. Due to the way many of these tools function, novel malware hiding methods can still go undetected.
Model poisoning
Key question: How do you ensure your model has not been maliciously modified?
NIST AML Mappings: NISTAML.011, .026, .051
Model poisoning refers to malicious modifications to an ML model. This can take various forms:
- availability poisoning – degrades model performance.
- targeted poisoning – adds bias or degrades specific aspects or categories of data.
- backdoor poisoning – embeds a hidden trigger that cause the model to behave abnormally when activated, while performing normally otherwise.
Due to the complex structure of ML models, these modifications can be hard to detect before testing.
Model-based malware embedding (stegomalware)
Key question: How do you ensure a model is not hiding malware?
NIST AML Mappings: Part of NISTAML.05
Due to their relatively large size and complex structure, ML models can serve as hosts for malicious content that may avoid conventional malware detection. Malware can hide segments of code in the weights or metadata of a model that antivirus software may not typically check.
Loading or running a malicious model could allow reconstruction and execution of malicious code, even while maintaining the expected performance. Malicious actors can use this to stealthily run malware on a system or stage it for later use.
Evasion attacks
Key question: How do you design a model to be resilient to unusual inputs?
NIST AML Mappings: NISTAML.022, .025
Malicious actors can engineer inputs specifically designed to evade ML-based security mechanisms such as spam filters or intrusion detection systems. They often circumvent simple rule-based methods by changing their evasion techniques in response to new controls. This can cause a model to have unintended or malicious behaviour.
While some mitigations (such as monitoring and filtering inputs and outputs) can be implemented after deployment, steps should also be taken in the supply chain to improve general resilience to unusual inputs.
Mitigations
To mitigate risks in ML models, validate them at the point of retrieval and continuously monitor their performance.
Secure file formats
Secure AI and ML model formats help protect model content, saving and loading processes to avoid serialisation attacks and similar issues. Approaches include using non-executable weight formats or immutable graph formats, and applying safe loading flags such as weight-only. These methods rarely have significant downsides compared to conventional serialisation methods, so use them where possible.
Trusted and transparent sources
Like data, use models from trusted and reputable providers where possible. It is a simple yet effective way to reduce the risk of poisoned or maliciously modified models but does not eliminate all risk. Treat all content sourced from the web as untrusted.
Look for models with good transparency around important details such as training data categories and sources, safety evaluations and bias testing. Choose models with AIBOMs or equivalent documentation that list their training data sources and known limitations. This information can help inform model choice and expectations.
Performance testing
Initial and periodic testing should be performed to verify that model behaviour is as expected and consistent. Deviations from expected results can indicate issues, such as model poisoning or drift. This can be especially important for models that use continuous learning or fine-tuning where there is a higher risk of model drift or performance degradation. If using an existing standard or historical dataset for comparison against live results, adequate steps should be taken to ensure its integrity. Any comparison would be invalid if this reference data were also poisoned or otherwise compromised.
Performance and other security tests should be performed initially and throughout the model’s lifecycle as testing methods and the model itself can change over time.
Model explainability and transparency
Explainable and transparent models help users understand how AI systems work and build trust in their outputs. These aspects can be improved in several ways. One technique is model simplification, which reduces structural complexity and makes the model logic easier to understand. Another is model visualisation, using graphs and diagrams to illustrate how the model is structured. Feature-importance analysis can also be used to highlight which aspects of data have the greatest influence on the model’s outputs. Where possible, employ models with inbuilt mechanisms to report internal processes and decision-making such as can be found in reasoning LLMs.
Improving explainability and trust also gives better insight into a model’s integrity and security, making it easier to identify issues like model poisoning. While this field is still evolving, it already benefits certain applications and is likely to become a key tool for AI systems as methods mature.
Model verification and provenance
Like AI data, use verification methods (such as checksums, hashes and digital signatures) to check the integrity of an ML model. For more information, refer to the section on Data verification and provenance under AI data.
Maintain a model registry of specific model versions that have been verified, tested, and approved for use. Verify and test each new version prior to approval and use.
Ensemble methods
Ensemble methods involve training multiple different datasets (or subsets) and models on the same problem to compare performance and consistency. For more information, refer to the Ensemble methods section under AI data.
Reproducible builds
A reproducible build refers to when developers release the components of a model rather than the complete and compiled version, like a SBOM. By including build instructions with these components, users can create the desired final model and verify its integrity with a checksum or similar method.
By starting with the components instead of the full model, it is easier for users to interrogate the components, reducing the risk of poisoning or embedded malware. AI-specific extensions to existing SBOM, or purpose-built AI BOMs, are emerging and are expected to help with this approach in the future.
Input-output monitoring
Ongoing input and output monitoring can be used to detect unusual or undesired model behaviour. It can identify malicious use, such as evasion attacks or prompt injection, or output anomalies, such as sensitive information or outliers that could indicate backdoor poisoning. As with performance testing, steps should be taken to ensure the integrity of any comparison data so evaluations can be accurate.
Adversarial training
Adversarial training introduces an adversarial model during the training process to create or augment training data in unexpected ways. This data helps to test weaknesses in a model, which helps improve its resilience to unusual or malicious inputs.
By challenging the model, it strengthens its resistance to poisoned data and adversarial techniques after deployment. It can also improve a model’s overall performance.
Model refining
Similar to data, some techniques can be applied to a model to help reduce the risk of malicious modifications. Effective techniques include pruning, compression and fine-tuning. These methods alter or remove weights and neurons within a model, which can disrupt the structure of embedded malware, often at the cost of performance.
Applying these techniques before using models can help reduce some poisoning and malware-embedding attacks.
Security tools
As model risks become more widely known, both open-source and proprietary tools are emerging that can help mitigate some of them. Consider incorporating relevant ML-specific security tools into your cyber supply chain. Examples include model scanners to check for malicious content or stegomalware, and adversarial emulators to assess resilience to evasion and extraction attacks. Performance evaluators can also ensure models perform accurately and as expected.
A growing range of tools can now address many cyber security considerations for AI and ML noted in this guidance though their effectiveness can be hard to verify. Users should assess the reputability and effectiveness of tools for themselves before relying on them.
AI software
AI and ML software includes any software that is used throughout the AI life cycle. This could be development frameworks, associated libraries, and other software that enables the system.
As with conventional cyber security, additional components can add risk by increasing a system’s attack surface if they are not verified and securely implemented.
Risks
Key question: How do you ensure that associated software and dependencies are not introducing additional vulnerabilities?
AI systems often rely on numerous libraries and tools, which increases complexity and makes it harder to ensure all components are secure. Even widely used tools and components can pose serious security risks.
Conventional threats (such as typosquatting and name-confusion attacks) remain relevant and should be considered when assessing risk for AI software.
Case study: Vulnerabilities in AI libraries
Several major AI frameworks and libraries have been found to have critical vulnerabilities that could enable deserialization or similar attacks. When exploited, these flaws can lead to token theft or even full remote code execution. In 2024, supply chain compromise of a published AI library led to users unknowingly installing cryptocurrency mining malware when installing the package.
Mitigations
Organisations should continue to validate all components of AI and ML systems as they would with any other piece of software.
Validate the integrity of software with checksums, digital signatures or provenance tracking to ensure it has not been tampered with during retrieval. This could also include using reproducible builds to provide greater visibility into sub-components.
Analyse software with malware scanning, static and dynamic application testing to ensure it is free of malware and operates as expected. Audit software components and maintain an SBOM to ensure known vulnerabilities are mitigated at deployment and throughout the system’s life cycle.
Deploy software securely by adhering to best practices such as least-privilege to ensure software is given only the minimum required permissions for its task. Plan for ongoing monitoring and maintenance to keep software and dependencies up to date with security patches.
AI infrastructure and hardware
AI systems have unique infrastructure and hardware requirements that can affect their security. Beyond high computational demands, AI systems more heavily rely on a GPU or even AI-specific accelerator devices.
Risks
Key question: How do you minimise the attack surface introduced by AI infrastructure and hardware?
Most risks to ML infrastructure and hardware mirror those found in conventional cyber systems, as they use broadly similar hardware. AI-specific accelerator devices introduce additional attack surfaces through their drivers, firmware and related components.
As machine-to-machine systems become more common, new communication protocols are being introduced. These may add complexity to systems, such as routing and firewall management.
Mitigations
Treat infrastructure and hardware for AI systems like any other integrated technology. Follow existing security processes and practices to ensure hardware is free from malicious content and appropriately segmented within the network. This includes enforcing signed drivers and firmware, enabling verified boot, logging, and using separate management networks with their own authentication and auditing.
Third-party services
Due to the resource demands and complexity of AI systems, third-party services are often used to support some or all components and services. This can range from using cloud-based data hosting, to developing and managing the entire AI system.
While third-party providers offer many advantages, they also come with additional risks and responsibilities.
Risks
Key question: How do you avoid risk from a third party’s poor cyber hygiene?
Third-party providers of AI and ML tools, platforms and services can introduce vulnerabilities into an organisation's supply chain. These vulnerabilities may be from their own supply chain, a cyber incident, or shared resources with other tenants.
Malicious actors can potentially use these vectors to compromise customers' systems if they are not well managed. Working with third parties can also create ambiguity around shared responsibilities that could lead to gaps in cyber security.
Mitigations
Perform thorough assessments of third party vendors, including reviewing their security practices, vulnerability management processes, and overall track record. Ongoing monitoring should ensure adherence to strong security practices.
Vendors should demonstrate a commitment to security and transparency across their products, services, systems and supply chain. Consider cyber security requirements for third parties early and include them in contracts. This could include restricting use of your data for training, defining separate cloud residencies, granting audit rights, setting continuity requirements, and establishing shared responsibility models.
Conclusion
AI and ML can introduce distinct cyber security challenges to a supply chain, particularly due to their reliance on a complex ecosystem of models, data, libraries and cloud infrastructure. Without proper safeguards, threats like poisoned data, backdoors and malicious code can compromise these systems and lead to further exploitation.
To mitigate these risks, organisations deploying or developing AI and ML systems and components should adopt secure supply chain practices that address the unique risks these systems introduce. By combining cyber security best practices and the mitigations in this guidance as part of a defence-in-depth strategy, organisations can reduce their attack surface and strengthen the overall resilience of their AI and ML supply chain.
More information
- Artificial intelligence
- Deploying AI systems securely
- Principles for the secure integration of Artificial Intelligence in Operational Technology
- Managing cyber supply chains
- Choosing Secure and Verifiable Technologies
- US CSCC Supplier, Products, and Services Threat Evaluation Report (to include Artificial Intelligence Risks and Mitigations)
- ANSSI Building trust in AI through a cyber risk-based approach
- CSA Guidelines and Companion Guide on Securing AI Systems
- MITRE ATLAS Matrix
- NIST AI Risk Management Framework
- CycloneDX Machine Learning Bill of Materials
- OWASP Machine Learning Security Top Ten