Cloudflare’s AI Safety for Apps detects and mitigates threats to AI-powered purposes. Right now, we’re asserting that it’s typically obtainable.
We’re transport with new capabilities like detection for customized subjects, and we’re making AI endpoint discovery free for each Cloudflare buyer—together with these on Free, Professional, and Enterprise plans—to offer everybody visibility into the place AI is deployed throughout their Web-facing apps.
We’re additionally asserting an expanded collaboration with IBM, which has chosen Cloudflare to ship AI safety to its cloud clients. And we’re partnering with Wiz to offer mutual clients a unified view of their AI safety posture.
A brand new type of assault floor
Conventional net purposes have outlined operations: verify a financial institution stability, make a switch. You may write deterministic guidelines to safe these interactions.
AI-powered purposes and brokers are completely different. They settle for pure language and generate unpredictable responses. There isn’t any mounted set of operations to permit or deny, as a result of the inputs and outputs are probabilistic. Attackers can manipulate giant language fashions to take unauthorized actions or leak delicate information. Immediate injection, delicate data disclosure, and unbounded consumption are only a few of the dangers cataloged within the OWASP High 10 for LLM Purposes.
These dangers escalate as AI purposes change into brokers. When an AI beneficial properties entry to instrument calls—processing refunds, modifying accounts, offering reductions, or accessing buyer information—a single malicious immediate turns into a direct safety incident.
Prospects inform us what they’re up in opposition to. “Most of Newfold Digital’s groups are placing in their very own Generative AI safeguards, however everyone is innovating so shortly that there are inevitably going to be some gaps finally,” says Rick Radinger, Principal Techniques Architect at Newfold Digital, which operates Bluehost, HostGator, and Area.com.
What AI Safety for Apps does
We constructed AI Safety for Apps to deal with this. It sits in entrance of your AI-powered purposes, whether or not you are utilizing a third-party mannequin or internet hosting your individual, as a part of Cloudflare’s reverse proxy. It helps you (1) uncover AI-powered apps throughout your net property, (2) detect malicious or off-policy conduct to these endpoints, and (3) mitigate threats through the acquainted WAF rule builder.
Discovery — now free for everybody
Earlier than you possibly can defend your LLM-powered purposes, you might want to know the place they’re getting used. We regularly hear from safety groups who don’t have a whole image of AI deployments throughout their apps, particularly because the LLM market evolves and builders swap out fashions and suppliers.
AI Safety for Apps mechanically identifies LLM-powered endpoints throughout your net properties, no matter the place they’re hosted or what the mannequin is. Beginning immediately, this functionality is free for each Cloudflare buyer, together with Free, Professional, and Enterprise plans.
Cloudflare’s dashboard web page of net property, displaying 2 instance endpoints labelled as cf-llm
Discovering these endpoints mechanically requires greater than matching frequent path patterns like /chat/completions. Many AI-powered purposes do not have a chat interface: suppose product search, property valuation instruments, or advice engines. We constructed a detection system that appears at how endpoints behave, not what they’re referred to as. To confidently establish AI-powered endpoints, adequate legitimate site visitors is required.
AI-powered endpoints which have been found can be seen underneath Safety → Net Belongings, labeled as cf-llm. For purchasers on a Free plan, endpoint discovery is initiated whenever you first navigate to the Discovery web page. For purchasers on a paid plan, discovery happens mechanically within the background on a recurring foundation. In case your AI-powered endpoints have been found, you possibly can overview them instantly.
AI Safety for Apps detections comply with the always-on method for site visitors to your AI-powered endpoints. Every immediate is run by a number of detection modules for immediate injection, PII publicity, and delicate or poisonous subjects. The outcomes—whether or not the immediate was malicious or not—are hooked up as metadata you should utilize in customized WAF guidelines to implement your insurance policies. We’re constantly exploring methods to leverage our world community, which sees site visitors from roughly 20% of the net, to establish new assault patterns throughout thousands and thousands of web sites earlier than they attain yours.
New in GA: Customized subjects detection
The product ships with built-in detection for frequent threats: immediate injections, PII extraction, and poisonous subjects. However each enterprise has its personal definition of what is off-limits. A monetary providers firm would possibly must detect discussions of particular securities. A healthcare firm would possibly must flag conversations that contact on affected person information. A retailer would possibly wish to know when clients are asking about competitor merchandise.
The brand new customized subjects function permits you to outline these classes. You specify the subject, we examine the immediate and output a relevance rating that you should utilize to log, block, or deal with nevertheless you resolve. Our aim is to construct an extensible instrument that flexes to your use instances.
Immediate relevance rating within AI Safety for Apps
AI Safety for Apps enforces guardrails earlier than unsafe prompts can attain your infrastructure. To run detections precisely and supply real-time safety, we first must establish the immediate throughout the request payload. Prompts can stay anyplace in a request physique, and completely different LLM suppliers construction their APIs in another way. OpenAI and most suppliers use $.messages[*].content material for chat completions. Anthropic’s batch API nests prompts inside $.requests[*].params.messages[*].content material. Your customized property valuation instrument would possibly use $.property_description.
Out of the field, we help the usual codecs utilized by OpenAI, Anthropic, Google Gemini, Mistral, Cohere, xAI, DeepSeek, and others. Once we cannot match a recognized sample, we apply a default-secure posture and run detection on the complete request physique. This may introduce false positives when the payload comprises fields which might be delicate however do not feed on to an AI mannequin, for instance, a $.customer_name subject alongside the precise immediate would possibly set off PII detection unnecessarily.
Quickly, you’ll outline your individual JSONPath expressions to inform us precisely the place to search out the immediate. This can cut back false positives and result in extra correct detections. We’re additionally constructing a prompt-learning functionality that may mechanically adapt to your software’s construction over time.
As soon as a risk is recognized and scored, you possibly can block it, log it, or ship customized responses, utilizing the identical WAF guidelines engine you already use for the remainder of your software safety. The facility of Cloudflare’s shared platform is that you could mix AI-specific alerts with the whole lot else we learn about a request, represented by a whole lot of fields obtainable within the WAF. A immediate injection try is suspicious. A immediate injection try from an IP that’s been probing your login web page, utilizing a browser fingerprint related to earlier assaults, and rotating by a botnet is a unique story. Level options that solely see the AI layer can’t make these connections.
This unified safety layer is strictly what they want at Newfold Digital to find, label, and defend AI endpoints, says Radinger: “We look ahead to utilizing it throughout all these initiatives to function a fail-safe.”
AI Safety for Purposes can even be obtainable by Cloudflare’s rising ecosystem, together with by integration with IBM Cloud. Via IBM Cloud Web Companies (CIS), finish customers can already procure superior software safety options and handle them immediately by their IBM Cloud account.
We’re additionally partnering with Wiz to attach AI Safety for Purposes with Wiz AI Safety, giving mutual clients a unified view of their AI safety posture, from mannequin and agent discovery within the cloud to application-layer guardrails on the edge.
AI Safety for Apps is on the market now for Cloudflare’s Enterprise clients. Contact your account group to get began, or see the product in motion with a self-guided tour.
In the event you’re on a Free, Professional, or Marketing strategy, you should utilize AI endpoint discovery immediately. Log in to your dashboard and navigate to Safety → Net Belongings to see which endpoints we have recognized. Maintain a watch out — we plan to make all AI Safety for Apps capabilities obtainable for patrons on all plans quickly.
For configuration particulars, see our documentation.



