For Risk & Compliance Officers4 min read

Your AI Supply Chain Is the Biggest Security Risk You're Not Managing

Researchers found 100+ weaponized AI models on a major public hub, and 83% of enterprises have no automated defenses.

The Problem

In February 2024, JFrog security researchers found more than 100 malicious AI models on Hugging Face, one of the most popular public repositories for machine learning. These models were not simply broken. They were weaponized. When a developer downloaded one and loaded it with a standard command, the model silently opened a backdoor — a reverse shell — to a remote server. The attacker immediately gained access to that developer's system.

One specific model, uploaded by a user called "baller423," used a common file format called pickle to inject code during the loading process. The moment a data scientist ran the standard torch.load() command, malicious code executed automatically. It connected to an IP address belonging to the Korea Research Environment Open Network.

This is not a theoretical risk. It is happening right now on the platforms your teams likely use every day. If your organization pulls AI models from public sources — and most do — you are importing binary files that no human can read or review by eye. Unlike traditional software code, AI model weights are opaque structures. Malicious behavior hides inside millions of numerical parameters. Your existing code review processes simply cannot catch this. The AI supply chain is now the most vulnerable and least governed part of your technology stack.

Why This Matters to Your Business

The financial and regulatory exposure here is real and growing. Consider these numbers from the whitepaper's research:

  • 83% of organizations lack automated AI security controls. Only 17% have any automated defenses in place. The rest are, in the researchers' framing, "operating blind."
  • Shadow AI — unauthorized use of AI tools by employees — accounts for 90% of enterprise AI usage. Your IT and security teams likely cannot see most of the AI activity in your organization right now.
  • Incidents involving unvetted AI tools increase the cost of a data breach by an average of $670,000. That is the premium you pay for ungoverned AI adoption.
  • Only 12% of enterprises have implemented AI governance with real technical controls. Another 56% say they are "ready" but lack the enforcement mechanisms to back it up.
  • 86% of organizations have no visibility into how AI moves data internally. If you cannot see the data flows, you cannot protect them or prove compliance.

For your board and your regulators, the question is straightforward. Can you show an auditor exactly which AI models are running in your environment, where they came from, and what data they touch? If you cannot, you face compliance gaps under emerging frameworks like NIST AI 100-2. You also face the reputational damage of a breach you never saw coming — because the attack vector was an AI model nobody vetted.

What's Actually Happening Under the Hood

To understand why this threat is so hard to catch, think of an AI model like a sealed recipe box. Traditional software is like a recipe written in plain text — you can read every instruction and spot something dangerous. An AI model, by contrast, is a locked box with millions of tiny dials inside. The dials produce outputs, but you cannot open the box and read the instructions. A poisoned model looks and performs exactly like a clean one — until a specific trigger activates the hidden behavior.

This is called data poisoning, and the math behind it is alarming. Researchers from Anthropic and NVIDIA's AI Red Team showed that poisoning just 0.00016% of training data — roughly 250 documents out of a massive dataset — permanently compromises a 13-billion parameter model. Once the poisoned data is baked into the model's weights, adding more clean data does not fix it. The backdoor is permanent.

These poisoned models act as "sleeper agents." They pass every standard test and benchmark. They behave normally in your QA environment. But when they encounter a specific trigger — a unique text string, a particular image pattern, even a bit-level input manipulation — the model switches to its malicious mode. That could mean bypassing authentication, exfiltrating sensitive data, or generating harmful code that your downstream systems execute automatically.

The danger extends beyond the models themselves. Even the security scanners meant to protect you are failing. JFrog found three zero-day vulnerabilities in PickleScan, a widely used tool for vetting AI models. Attackers could manipulate file extensions to make a compromised model appear "safe." Research suggests up to 96% of current scanner alerts are false positives, which desensitizes your security team to real threats.

What Works (And What Doesn't)

Let's start with three common approaches that fall short:

  • Relying on public repository trust scores. Hugging Face is not a vetted software vendor. The 100+ malicious models found there prove that popularity and availability do not equal safety.
  • Writing AI governance policies without automated enforcement. A policy document does not stop a developer from downloading an unvetted model at 2 AM. Without technical barriers, 77% of employees will continue pasting proprietary data into public AI tools.
  • Running standard code scanners on AI model files. Traditional static analysis tools lack the context to inspect model weights. They cannot detect backdoors hidden in numerical parameters or malicious templates embedded in model metadata.

What does work is treating every AI model as potentially malicious executable code. Here is the architecture principle in three steps:

  1. Input — Create a Machine Learning Bill of Materials (ML-BOM). Before any model enters your environment, generate a tamper-proof record of its origin, training data, framework dependencies, and a cryptographic signature. Think of this as a chain-of-custody document for every AI artifact. Standards like CycloneDX and SPDX 3.0 now support AI-specific profiles for exactly this purpose. Your data provenance and traceability practices form the foundation here.

  2. Processing — Enforce cryptographic verification at load time. Your inference server — the system that runs AI models — should include an "Admission Controller" that checks every model's digital signature against a trusted registry before it loads into memory. This means only models signed by your organization's own keys can execute. Use Hardware Security Modules to manage those keys. Ban the pickle format entirely in favor of SafeTensors, a format that does not permit executable code.

  3. Output — Monitor model behavior continuously at runtime. Compare model outputs against clean validation baselines to detect drift or anomalies that signal backdoor activation. Apply input sanitization layers that rephrase queries before they reach the model, disrupting carefully crafted trigger payloads. Rate-limit queries to prevent model extraction attacks.

The audit trail advantage is what makes this approach valuable to your compliance team. Every model has a signed ML-BOM. Every deployment has a verified signature chain. Every inference has monitored outputs. When your regulator or auditor asks "how do you know this AI model is safe?" you can point to cryptographic proof — not a policy document.

For organizations in AI security and resilience, this is not optional anymore. The convergence of AI security and software supply chain security means your existing CI/CD pipeline protections must extend to cover model artifacts. If your model is secure but the Python library running it is compromised, your system is still breached. A unified approach to security assessment and hardening across both software and AI assets is the only way to close this gap.

Read the full technical analysis for detailed implementation guidance, or explore the interactive version to see how these controls map to your specific environment.

Key Takeaways

  • JFrog researchers found 100+ malicious AI models on Hugging Face in 2024, many containing backdoors that execute code the moment a developer loads them.
  • Poisoning just 0.00016% of training data permanently compromises a 13-billion parameter AI model — and the backdoor survives additional clean training.
  • 83% of enterprises lack automated AI security controls, and 90% of enterprise AI usage happens outside IT's visibility as Shadow AI.
  • Unvetted AI tools add an average of $670,000 to the cost of a data breach.
  • Cryptographically signed model artifacts with a Machine Learning Bill of Materials provide the auditable proof trail regulators increasingly demand.

The Bottom Line

Your AI supply chain is almost certainly less secure than your traditional software supply chain, and attackers know it. The combination of unvetted public models, invisible Shadow AI, and security scanners riddled with false positives creates an exposure most enterprises cannot even measure yet. Ask your AI vendor: can you show us the cryptographic signature and full provenance chain for every model running in our environment right now?

FAQ

Frequently Asked Questions

Can AI models from Hugging Face be trusted for enterprise use?

Not without vetting. In February 2024, JFrog researchers found more than 100 malicious models on Hugging Face containing backdoors for remote code execution. One model opened a reverse shell to a remote server the moment it was loaded. Enterprises should treat every public model as potentially malicious and enforce cryptographic verification before deployment.

What is AI data poisoning and how much data does it take?

AI data poisoning is when attackers manipulate training data to embed hidden backdoors in a model. Research from Anthropic and NVIDIA showed that poisoning just 0.00016% of training data — about 250 documents — permanently compromises a 13-billion parameter model. The backdoor survives additional clean training and passes standard testing benchmarks.

How do I protect my company from Shadow AI risks?

Shadow AI accounts for 90% of enterprise AI usage and increases data breach costs by an average of $670,000. Protection requires a centralized AI asset registry, automated vetting pipelines for any model entering your environment, and mandatory Machine Learning Bills of Materials that document model provenance. Policy documents alone are not enough — you need automated technical controls.

Build Your AI with Confidence.

Partner with a team that has deep experience in building the next generation of enterprise AI. Let us help you design, build, and deploy an AI strategy you can trust.

Veriprajna Deep Tech Consultancy specializes in building safety-critical AI systems for healthcare, finance, and regulatory domains. Our architectures are validated against established protocols with comprehensive compliance documentation.