Skip to main content
Mallory
Mallory

Insecure Public Exposure of Self-Hosted AI Infrastructure (Ollama and MCP Servers)

shadow aipublic exposureunauthenticated accessinference serversmisconfigurationmcpdata lossmodel context protocolaudit trailsaccess controlsarbitrary command executionai agentsllm gatewaysai governanceasset inventory
Updated February 7, 2026 at 01:03 AM2 sources
Insecure Public Exposure of Self-Hosted AI Infrastructure (Ollama and MCP Servers)

Get Ahead of Threats Like This

Know if you're exposed — before adversaries strike.

Security researchers and media reporting highlighted widespread public exposure of self-hosted AI infrastructure caused by rushed, poorly governed deployments. Reporting cited 14,000+ internet-accessible Ollama inference servers, with one analysis estimating ~20% hosting models susceptible to unauthorized access, and separate findings identifying 10,000+ Ollama servers exposed without any authentication—often due to developers binding services to all interfaces or standing up local inference/gateway components (e.g., LiteLLM, vLLM) outside normal asset inventories. The net effect is “shadow AI” that creates material blind spots for security teams and increases the likelihood of unauthorized model access, data exposure, and abuse of internal AI services.

In parallel, enterprise adoption of Model Context Protocol (MCP) servers—which bridge LLMs to internal tools and data—has introduced similar exposure risk when deployed without access controls. Guidance and analysis noted that MCP, introduced as an open standard without native role restrictions, leaves security implementation to operators; researchers reportedly identified nearly 2,000 MCP servers on the open web with no security controls, increasing risk of unauthorized access, data loss, and potentially arbitrary command execution via overly privileged integrations. A vendor announcement positioned an AI-agent governance platform (MintMCP) as a response to these visibility and control gaps (audit trails, policy enforcement, access controls), but it primarily serves as product marketing rather than independent incident reporting.

Related Stories

Security Exposure and Threat Landscape for Model Context Protocol (MCP) Servers

Security researchers evaluated the risks associated with deploying Model Context Protocol (MCP) servers, which enable AI systems like ChatGPT to interact with external tools and data. One investigation used the GitHub MCP server in conjunction with OpenAI's Codex to analyze code, identify security issues, and propose fixes, highlighting how AI agents can streamline code review and vulnerability management. The study also explored whether AI-driven code analysis could be manipulated to conceal security flaws, emphasizing the importance of context and transparency in automated security workflows. Separately, honeypots simulating MCP server deployments were exposed to the internet to assess real-world attack activity. These honeypots, configured with varying authentication levels, were quickly discovered by internet scanners but did not experience targeted exploitation or MCP-specific attacks. The only notable incident was a controlled proof-of-concept prompt-hijacking flaw in a custom MCP build, which was not observed in the wild. The findings suggest that, while MCP servers are rapidly indexed by threat actors, current risks stem primarily from implementation errors rather than active targeting, underscoring the need for secure deployment practices and ongoing monitoring as MCP adoption grows.

4 months ago
Security Risks and Attacks Targeting Large Language Model (LLM) Services and AI Integration Protocols

Security Risks and Attacks Targeting Large Language Model (LLM) Services and AI Integration Protocols

Attackers have increasingly targeted exposed large language model (LLM) services and the protocols that enable their integration, such as the Model Context Protocol (MCP). GreyNoise researchers observed nearly 100,000 attack sessions against public LLM endpoints, with campaigns probing for misconfigured proxies and server-side request forgery vulnerabilities to map the expanding AI attack surface. These attacks, which included methodical enumeration of OpenAI-compatible and Google Gemini endpoints, highlight the growing risk as enterprises move LLM deployments from experimental to production environments. Security experts warn that such enumeration efforts are likely precursors to more serious exploitation, emphasizing the need for organizations to secure exposed LLM endpoints and monitor for abnormal access patterns. The Model Context Protocol (MCP), designed to facilitate seamless integration between LLMs and external tools, has also been identified as a double-edged sword. While MCP enables powerful automation and workflow enhancements, it extends the attack surface by embedding trust in external products and services, making it susceptible to exploitation by adversaries who manipulate context layers and metadata. Security leaders, such as Block's CISO, stress the importance of applying least-privilege principles and rigorous red-teaming to AI agents and integration protocols, recognizing that both human and machine actors can introduce significant risks. As LLMs and AI agents become ubiquitous in enterprise environments, organizations must adapt their security frameworks to address these novel attack vectors and integration challenges.

2 months ago
Model Context Protocol (MCP) Security Risks From Untrusted Tool Servers and Verifiability Gaps

Model Context Protocol (MCP) Security Risks From Untrusted Tool Servers and Verifiability Gaps

Security researchers warned that the *Model Context Protocol (MCP)*—used to let AI assistants connect to local tools and enterprise SaaS data—creates a significant attack surface when organizations install or authorize MCP “servers” and tool integrations. Praetorian highlighted that **locally hosted MCP servers run with the user’s privileges** and can therefore execute arbitrary commands, access local files, install malware, and exfiltrate data while masquerading as legitimate productivity tooling; it also described **“MCP server chaining,”** where a malicious local MCP server abuses data and actions flowing through a trusted remote integration (e.g., Slack/Google Drive) without needing to compromise the official provider. Separately, Gopher Security emphasized a **trust and auditability gap** in MCP deployments: standard logging for remote tool execution can be incomplete or tampered with, and organizations often cannot prove what code ran or what parameters were used inside a remote “black box” execution environment. The post described “puppet”/interception-style scenarios where an attacker could alter an MCP request (e.g., changing tool-call parameters to trigger data exfiltration or unauthorized actions) while returning plausible “success” responses, and proposed cryptographic approaches (e.g., **zero-knowledge proofs**) to make MCP tool execution verifiable rather than relying on mutable logs.

3 weeks ago

Get Ahead of Threats Like This

Mallory continuously monitors global threat intelligence and correlates it with your attack surface. Know if you're exposed — before adversaries strike.

Insecure Public Exposure of Self-Hosted AI Infrastructure (Ollama and MCP Servers) | Mallory