Security researchers have identified critical vulnerabilities in Hugging Face’s open source AI model platform that could allow attackers to hijack model outputs and distribute malware through fake repositories. The flaws affect popular model formats including SafeTensors, ONNX, and GGUF, potentially impacting thousands of AI developers and organizations using locally-run models.
Tokenizer Vulnerability Enables Man-in-the-Middle Attacks
HiddenLayer security researchers discovered that Hugging Face’s tokenizer system can be weaponized through manipulation of a single JSON file. According to HiddenLayer’s analysis, attackers can implement man-in-the-middle attacks by modifying the tokenizer layer that converts AI model outputs into human-readable text.
The attack vector works by intercepting tool call arguments and redirecting URL tokens through attacker-controlled infrastructure. This gives threat actors “visibility into every URL the model accesses, API parameters, and any credentials embedded in those requests,” explained HiddenLayer researcher Divyanshu Divyanshu in a blog post released Monday.
The vulnerability specifically affects locally-run models across three major formats: SafeTensors (Hugging Face’s de facto standard), ONNX, and GGUF. Models accessed through Hugging Face’s cloud-based Inference API remain unaffected since the attack requires modifying local files. The flaw could also impact other open source model platforms including LlamaCPP and Ollama.
Fake Repository Campaign Targets AI Developers
Separately, security researchers have identified a supply chain attack involving fake OpenAI repositories on Hugging Face that distribute infostealer malware. The campaign specifically targets AI developers and organizations building applications with open source models.
The malicious repositories masquerade as legitimate OpenAI model releases, exploiting the trust developers place in the Hugging Face ecosystem. When developers download these fake models, they unknowingly install malware designed to steal credentials, API keys, and other sensitive information from their development environments.
This attack highlights the broader security challenges facing the open source AI ecosystem, where the decentralized nature of model distribution creates opportunities for malicious actors to insert compromised packages into the supply chain.
Impact on Open Source AI Ecosystem
These security revelations come at a critical time for open source AI development. Hugging Face hosts over 500,000 models and has become the de facto repository for researchers and companies building AI applications. Major models from Meta’s Llama family, Mistral AI, and hundreds of other organizations rely on the platform for distribution.
The tokenizer vulnerability is particularly concerning because it affects the core functionality that makes AI models usable. Every interaction between users and locally-run models passes through the tokenizer system, creating a chokepoint that attackers can exploit to monitor or manipulate AI outputs.
For enterprise users, these vulnerabilities pose significant risks to data privacy and operational security. Organizations using open source models for sensitive applications like document analysis, code generation, or customer service could inadvertently expose confidential information to attackers.
Platform Security Response and Mitigation
Hugging Face has not yet responded to requests for comment regarding the tokenizer vulnerability. The company’s security team typically issues patches and guidance through their official channels when vulnerabilities are disclosed.
Developers can mitigate risks by implementing several security practices. For the tokenizer vulnerability, organizations should audit local model files before deployment and monitor network traffic from AI applications for unexpected connections. Regular updates to model files and dependencies can help prevent exploitation of known vulnerabilities.
To avoid fake repository attacks, developers should verify model authenticity through official channels and check repository metadata for signs of legitimacy, including contributor history and community engagement metrics.
What This Means
These security discoveries underscore the growing pains of the open source AI ecosystem as it scales from research tool to enterprise infrastructure. While open source models offer significant advantages in cost, customization, and transparency compared to proprietary alternatives, they also inherit the security challenges common to all open source software.
The timing is particularly significant as enterprises accelerate AI adoption and increasingly rely on open source models for production workloads. Organizations must balance the benefits of open source AI with robust security practices, including thorough vetting of model sources and implementation of monitoring systems.
For the broader AI industry, these vulnerabilities highlight the need for standardized security frameworks specifically designed for AI model distribution and deployment. As the ecosystem matures, expect to see enhanced security tooling and best practices emerge to address these challenges.
FAQ
What is the tokenizer vulnerability in Hugging Face models?
The vulnerability allows attackers to modify a JSON file in the tokenizer system to intercept and redirect model communications. This enables man-in-the-middle attacks that can expose URLs, API parameters, and embedded credentials from locally-run AI models.
Which AI models are affected by these security issues?
The tokenizer vulnerability affects locally-run models in SafeTensors, ONNX, and GGUF formats on Hugging Face and potentially other platforms like LlamaCPP and Ollama. Cloud-hosted models accessed through APIs are not affected. The fake repository attacks can target any open source model distributed through Hugging Face.
How can developers protect against these AI security threats?
Developers should verify model sources through official channels, audit local model files before deployment, monitor network traffic from AI applications, and maintain updated dependencies. Organizations should implement security scanning for AI model files similar to traditional software security practices.
Related news
- Developers can now debug and evaluate AI agents locally with Raindrop’s open source tool Workshop – VentureBeat
- ThreatsDay Bulletin: PAN-OS RCE, Mythos cURL Bug, AI Tokenizer Attacks, and 10+ Stories – The Hacker News
- Japanese text opens new legal tech chapter – Financial Times Tech
Sources
- Fine-Tuning Your First Large Language Model (LLM) with PyTorch and Hugging Face – HuggingFace Blog
- Hugging Face Packages Weaponized With a Single File Tweak – Dark Reading
- Perceptron Mk1 shocks with highly performant video analysis AI model 80-90% cheaper than Anthropic, OpenAI & Google – VentureBeat
- So you’ve heard these AI terms and nodded along; let’s fix that – TechCrunch
- Supply Chain Attack: Fake OpenAI Repository on Hugging Face Distributes Infostealer Malware Targeting Developers and AI Tools – Rescana – Google News – AI Tools






