Last update:2026-03-25 18:16:53
If you need to manage specific automated programs separately on your website, you can use Custom Bots. This feature lets you identify bots based on characteristics such as User-Agent, request headers, IP address, and fingerprints, and then apply actions such as skip, deny, or log. This enables more granular traffic control for Bots.
Note: Custom Bots are evaluated after IP/Region Blocking, Custom Rules, Whitelist, DDoS Protection, and Rate Limiting. You can think of this feature as access control or custom rule configuration within the Bot Management module.
In most cases, Custom Bots are used to prevent certain automation tools, such as a partner’s self-service ordering tool, from being identified as bots and blocked. However, if those tools are abused to launch attacks such as DDoS, the DDoS Protection policy will still detect and mitigate them.
You can log, deny, or skip Custom Bot traffic.
| Action | Description |
|---|---|
| Log | Only log this type of request and forward it as usual. |
| Deny | Deny the request and respond with 403. |
| Skip | Log the request and skip all subsequent Bot policy inspections, but other protections such as WAF and API Security will still be enforced. |
, and select the Domain where you want to apply this rule.SEO Traffic Protection: Allows major search engines, such as Google, to access your site smoothly, while blocking other crawlers that consume bandwidth without providing business value. For example:
Set the Public Bots - Search Engine category to block

Add Google to the Custom Bots for skip
