How to Block PerplexityBot
Complete guide to blocking PerplexityBot (Perplexity) from crawling your website using robots.txt, server configuration, and Switch workflows.
Should You Block PerplexityBot?
Blocking PerplexityBot prevents your content from appearing in Perplexity's AI-generated answers. Each visit from this agent represents a real user asking about your content.
Consider allowing PerplexityBot for visibility, or use Switch to serve agent-optimized markdown content instead of blocking entirely.
Blocking Methods
1robots.txt
High for cooperative crawlersAdd a Disallow rule for PerplexityBot's user-agent string in your robots.txt file. This is the standard, cooperative method that well-behaved crawlers respect.
2Server-side UA filtering
HighConfigure your web server (nginx, Apache, Cloudflare) to reject requests matching PerplexityBot's user-agent patterns. This blocks at the network level before your application processes the request.
3Switch Journey Workflows
Highest — granular, real-time controlCreate a custom journey in Switch that detects PerplexityBot and routes it to a block action, challenge, redirect, or modified content — without touching your server configuration.
robots.txt — Block PerplexityBot
Add the following to your robots.txt file (at the root of your domain) to block PerplexityBot:
User-agent: PerplexityBot Disallow: / User-agent: perplexity Disallow: /
robots.txt — Allow with Restrictions
Alternatively, allow PerplexityBot on most pages while blocking specific directories:
User-agent: PerplexityBot Disallow: /private/ Allow: / User-agent: perplexity Disallow: /private/ Allow: /
PerplexityBot User-Agent Strings
Use these patterns to identify PerplexityBot in your server logs or firewall rules:
Frequently Asked Questions
Does blocking PerplexityBot affect my Google search rankings?
No. Blocking PerplexityBot does not affect your Google search rankings. Only blocking Googlebot impacts Google Search visibility.
Does PerplexityBot respect robots.txt?
Yes, PerplexityBot respects robots.txt directives. Adding a Disallow rule for its user-agent will prevent it from crawling blocked paths.
Can I allow PerplexityBot on some pages but not others?
Yes. Use robots.txt to disallow specific directories, or use Switch journey workflows for granular page-level control with conditional logic.
Go beyond robots.txt
Switch detects PerplexityBot in real-time and lets you build custom journey workflows — block, challenge, redirect, or serve modified content. No server changes required.
Get Started Free