{"id":130930,"date":"2025-10-16T08:55:16","date_gmt":"2025-10-16T12:55:16","guid":{"rendered":"https:\/\/www.inmotionhosting.com\/support\/?p=130930"},"modified":"2025-11-12T17:28:33","modified_gmt":"2025-11-12T22:28:33","slug":"how-to-use-robots-txt-to-block-crawlers","status":"publish","type":"post","link":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/","title":{"rendered":"How to Use robots.txt Disallow to Block Crawlers and Protect Site Performance"},"content":{"rendered":"<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"538\" src=\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-1024x538.png\" alt=\"How to Use robots.txt Hero Image\" class=\"wp-image-130931\" srcset=\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-1024x538.png 1024w, https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-300x158.png 300w, https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-768x403.png 768w, https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png 1200w\" sizes=\"auto, (min-width: 1360px) 876px, (min-width: 960px) calc(61.58vw + 51px), calc(100vw - 80px)\" \/><\/figure>\n\n\n<div class=\"wp-block-post-excerpt\"><p class=\"wp-block-post-excerpt__excerpt\">Crawlers drive search visibility, but they can also overload your server if left unchecked. This guide shows you how to use robots.txt to take control. Learn how to block low-value directories, protect essential assets that Google needs to render your site, manage crawl rates with delays, and defend against bandwidth-heavy AI bots. You\u2019ll get practical examples, real-world use cases, and best practices for combining robots.txt with server rules. Strategic crawler management improves site speed, reduces hosting costs, and ensures search engines focus on your most important content. <\/p><\/div>\n\n\n<p>Search engines discover your content by sending automated crawlers, often called bots or spiders, that scan your pages and index them. This process is essential for visibility in Google or Bing, but if left uncontrolled, bots can overrun your server and slow down performance.<\/p>\n\n\n\n<p>That\u2019s where the robots.txt file comes in. It\u2019s one of the simplest yet most powerful tools for directing how search engines interact with your website. With a few lines of text, you can tell compliant bots what to explore and what to leave alone.<\/p>\n\n\n\n<p>Whether you\u2019re running a <a href=\"https:\/\/www.inmotionhosting.com\/wordpress-hosting\">high-performance WordPress site<\/a>, managing multiple client websites, or scaling an e-commerce platform, controlling crawler access protects your infrastructure and keeps your site running at peak speed.<\/p>\n\n\n\n<p>However, not all crawlers play by the rules. A growing wave of AI-driven bots are consuming massive amounts of bandwidth without contributing to your SEO visibility. Knowing how to manage both traditional and modern crawlers is now part of maintaining your site\u2019s speed, stability, and search performance.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">What Robots.txt Really Does (and What It Doesn\u2019t)<\/h2>\n\n\n\n<p>Before you begin blocking or allowing crawlers, it\u2019s important to understand what this small file actually controls. Many site owners assume robots.txt is a universal shield, but it\u2019s more of a polite request that good bots usually respect. Understanding its limits helps you set realistic expectations and use it strategically.<\/p>\n\n\n\n<p>When a search engine visits your site, it first looks for a robots.txt file in the root directory (for example, https:\/\/example.com\/robots.txt). The file tells the crawler which parts of the site it can or cannot access.<\/p>\n\n\n\n<p>For example:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono\" style=\"font-size:.875rem;font-family:Code-Pro-JetBrains-Mono,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:1.25rem;--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: *\nDisallow: \/private\/<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2m-6 9l2 2 4-4\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #79B8FF\">*<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/private\/<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>This directive tells all bots to avoid the <code>\/private\/<\/code> directory.<\/p>\n\n\n\n<p>But it\u2019s important to understand what robots.txt does not do:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>It doesn\u2019t physically block access to a page; it simply requests that compliant bots avoid it.<\/li>\n\n\n\n<li>It\u2019s not a security tool, anyone can view your robots.txt file in a browser.<\/li>\n\n\n\n<li>It doesn\u2019t guarantee exclusion from search results; pages may still appear if other sites link to them.<\/li>\n<\/ul>\n\n\n\n<p>For total exclusion, combine robots.txt with a noindex meta tag or server-level access restrictions.<\/p>\n\n\n\n<p>When you see robots.txt as a guidance system rather than a gate, you\u2019ll start using it with more precision. The next step is knowing when crawl control helps your site, and when it might hurt your SEO.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img decoding=\"async\" src=\"https:\/\/www.inmotionhosting.com\/blog\/wp-content\/uploads\/2025\/10\/Permission-conflicts-6-1024x1024.png\" alt=\"Robots.txt Clarifications\" class=\"wp-image-81265\" style=\"width:525px;height:auto\"><\/figure>\n<\/div>\n\n\n<h2 class=\"wp-block-heading\">When You Should Block Crawlers<\/h2>\n\n\n\n<p>Knowing when to block bots can make your website faster and your crawl allocation more efficient. Search engines give each site a limited amount of crawling time. Essentially, how many pages they\u2019ll check during each visit. When you guide crawlers away from low-value pages, they can spend more time indexing the content that matters.<\/p>\n\n\n\n<p>There are legitimate reasons to <a href=\"https:\/\/www.webmasterworld.com\/webmaster\/5119367.htm\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">limit crawling<\/a>. Doing so can conserve resources, improve SEO efficiency, and protect non-public sections of your site.<\/p>\n\n\n\n<p><a href=\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-stop-search-engines-from-crawling-your-website\/\">Common Reasons to Block Crawlers:<\/a><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Private or administrative areas:<\/strong> \/wp-admin\/, \/cgi-bin\/, \/tmp\/<\/li>\n\n\n\n<li><strong>Temporary or duplicate environments:<\/strong> staging or test directories<\/li>\n\n\n\n<li><strong>Dynamic URLs that generate endless combinations:<\/strong> filtered search or parameter-based pages<\/li>\n\n\n\n<li><strong>Resource-heavy files:<\/strong> large PDFs, feeds, or scripts that don\u2019t need indexing<\/li>\n<\/ul>\n\n\n\n<p>Here\u2019s an example:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono\" style=\"font-size:.875rem;font-family:Code-Pro-JetBrains-Mono,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:1.25rem;--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: *\nDisallow: \/cgi-bin\/\nDisallow: \/private\/\nDisallow: \/tmp\/<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2m-6 9l2 2 4-4\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #79B8FF\">*<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/cgi-bin\/<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/private\/<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/tmp\/<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>This keeps crawlers focused on your high-value content, protecting your server resources while improving crawl efficiency.<\/p>\n\n\n\n<p>Blocking isn\u2019t just about defense, it\u2019s about clarity. The right disallow rules tell search engines exactly where to focus, helping them crawl your site faster and more effectively. But remember, blocking the wrong resources can do more harm than good.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">When Blocking Crawlers Hurts Your SEO<\/h2>\n\n\n\n<p>While blocking crawlers can protect performance, doing it carelessly can cause major SEO damage. It\u2019s easy to overlook which assets are essential for rendering your site, especially if you\u2019re managing multiple themes or plugins. When search engines can\u2019t see how your site actually loads, they may misjudge quality or relevance.<\/p>\n\n\n\n<p>Blocking crawlers without strategy can backfire. If essential assets (like CSS or JavaScript) are blocked, Google can\u2019t render your site correctly, which may affect rankings.<\/p>\n\n\n\n<p><strong>Avoid blocking:<\/strong><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>\/wp-content\/ (theme images, scripts, styles)<\/li>\n\n\n\n<li>\/wp-includes\/ (WordPress core functionality)<\/li>\n\n\n\n<li>Category or tag pages that provide navigation structure<\/li>\n<\/ul>\n\n\n\n<p><a href=\"https:\/\/developers.google.com\/search\/blog\/2014\/10\/updating-our-technical-webmaster\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">In 2015<\/a>, Google began rendering pages like a modern browser. If CSS or JS files are blocked, Google might see a broken layout, slow load time, or missing interactive elements. All of these can damage your <a href=\"https:\/\/www.inmotionhosting.com\/blog\/core-web-vitals\/\">Core Web Vitals<\/a> and ultimately hurt your visibility.<\/p>\n\n\n\n<p>A good rule of thumb is this: if a file affects what a human visitor sees or interacts with, don\u2019t block it. Once you know what to keep open, you can start fine-tuning how bots behave to protect your performance without hurting SEO.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">The Performance Side of Crawl Management<\/h2>\n\n\n\n<p>Performance and crawl control are closely connected. Every bot request uses server bandwidth, and too many at once can push your resources to the limit. Understanding this relationship helps you avoid technical slowdowns that look like SEO problems but are really traffic management issues.<\/p>\n\n\n\n<p>Uncontrolled crawler activity doesn\u2019t just clutter search results, it can overload your hosting account.<a href=\"https:\/\/www.inmotionhosting.com\/support\/website\/seo\/search-engine-crawlers-automated-robots-not-being-controlled\/\"> Even a few bots can generate the resource impact<\/a> of thousands of human visitors.<\/p>\n\n\n\n<p>Unlike people, bots don\u2019t pause between clicks. They request every linked page in rapid succession, which can:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Spike CPU and RAM usage<\/li>\n\n\n\n<li>Exhaust your bandwidth quota<\/li>\n\n\n\n<li>Trigger account throttling or even suspension on shared hosting plans<\/li>\n<\/ul>\n\n\n\n<p>A single misbehaving bot can cause temporary outages or performance degradation, especially when it crawls large archives or resource-heavy pages.<\/p>\n\n\n\n<p>Keeping performance steady isn\u2019t just about hardware. It\u2019s about setting healthy boundaries for automation. The next section shows how to slow crawlers down safely using the crawl-delay directive.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Controlling Crawl Rate with Crawl-Delay<\/h2>\n\n\n\n<p>Once you\u2019ve identified aggressive crawler behavior, you can manage it without outright blocking. The Crawl-delay directive gives you fine-tuned control over how often bots make requests. By pacing the crawl rate, you reduce stress on your server while maintaining accessibility for legitimate search engines.<\/p>\n\n\n\n<p>If your site has hundreds or thousands of pages, even legitimate bots like Googlebot can cause strain when they crawl too quickly. You can slow them down using the Crawl-delay directive.<\/p>\n\n\n\n<p>For example:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono\" style=\"font-size:.875rem;font-family:Code-Pro-JetBrains-Mono,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:1.25rem;--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: *\nCrawl-delay: 30<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2m-6 9l2 2 4-4\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #79B8FF\">*<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Crawl-delay:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #79B8FF\">30<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>This tells all bots to wait 30 seconds between requests.<\/p>\n\n\n\n<p>A 30-second crawl delay on a 1,000-page site allows indexing in about 8 hours instead of minutes, keeping your server responsive for real visitors.<\/p>\n\n\n\n<p>However, note that Google ignores Crawl-delay in robots.txt; it must be adjusted within Google Search Console\u2019s crawl settings instead.<\/p>\n\n\n\n<p>Crawl-delay is a balance between visibility and performance. When used correctly, it keeps your site fast while giving search engines the time they need to explore. Next, let\u2019s look at how to set up and manage this file in your hosting account.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">How to Create or Edit Robots.txt <\/h2>\n\n\n\n<p>Once you understand the logic behind your crawl rules, adding them to your website is straightforward. Most InMotion Hosting users can manage robots.txt through cPanel or an SEO plugin without touching the command line. The key is to keep the file accessible and accurate.<\/p>\n\n\n\n<p>You can manage your <code>robots.txt<\/code> file directly from your hosting account:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Access via File Manager (cPanel):<\/strong>\n<ul class=\"wp-block-list\">\n<li>Go to your site\u2019s root directory, usually <code>\/public_html\/<\/code>.<\/li>\n\n\n\n<li>If the file doesn\u2019t exist, create a new plain-text file named <code>robots.txt<\/code>.<\/li>\n<\/ul>\n<\/li>\n\n\n\n<li><strong>Edit the file:<\/strong>\n<ul class=\"wp-block-list\">\n<li>Open it in the built-in editor or use FTP to modify it.<\/li>\n\n\n\n<li>Paste in your desired rules, such as: <code>User-agent: * Disallow: \/private\/<\/code><\/li>\n<\/ul>\n<\/li>\n\n\n\n<li><strong>Save and test:<\/strong>\n<ul class=\"wp-block-list\">\n<li>Visit <code>https:\/\/yourdomain.com\/robots.txt<\/code> to confirm the file is live.<\/li>\n\n\n\n<li>Use Google Search Console\u2019s \u201crobots.txt Tester\u201d to validate your syntax.<\/li>\n<\/ul>\n<\/li>\n<\/ol>\n\n\n\n<p class=\"alert alert-info\">For WordPress users, SEO plugins like <strong>Yoast SEO<\/strong> or <strong>All in One SEO<\/strong> include tools for editing robots.txt safely from the dashboard.<\/p>\n\n\n\n<p>Maintaining an accurate <code>robots.txt<\/code> file gives you direct control over what bots see. But the landscape of crawlers is changing quickly, especially with AI systems entering the picture. Let\u2019s explore what that means for your performance strategy.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img decoding=\"async\" src=\"https:\/\/www.inmotionhosting.com\/blog\/wp-content\/uploads\/2025\/10\/Blue-and-White-Simple-Daily-Vlogger-YouTube-Banner-1-1024x576.png\" alt=\"How to Create or Edit Robots.txt\" class=\"wp-image-81267\" style=\"width:688px;height:auto\"><\/figure>\n<\/div>\n\n\n<h2 class=\"wp-block-heading\">Modern Crawl Challenges: AI and LLM Bots<\/h2>\n\n\n\n<p>The rise of AI crawlers has changed what \u201cbot management\u201d means. Traditional SEO bots helped your site appear in search results; AI bots extract massive amounts of text to train models. Managing this traffic is no longer optional, it\u2019s a key part of protecting your bandwidth and brand.<\/p>\n\n\n\n<p>A new class of crawlers (AI and Large Language Model (LLM) bots) has emerged. These bots, such as GPTBot, ClaudeBot, and PerplexityBot, harvest large amounts of web data to train AI systems.<\/p>\n\n\n\n<p><a href=\"https:\/\/www.inmotionhosting.com\/blog\/ai-crawlers-slowing-down-your-website\/\">AI crawlers can cause massive traffic spikes:<\/a><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>GPTBot generated <strong>569 million requests<\/strong> in one month on Vercel\u2019s network.<\/li>\n\n\n\n<li>One site owner reported <strong>30TB of bandwidth<\/strong> consumed by AI crawlers in a single month.<\/li>\n\n\n\n<li>Over <strong>35% of the top 1,000 websites<\/strong> now block GPTBot with robots.txt (PPC Land, 2024).<\/li>\n<\/ul>\n\n\n\n<p>Unlike Googlebot, AI crawlers often:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Ignore crawl delays or bandwidth-saving standards<\/li>\n\n\n\n<li>Request content in large bursts<\/li>\n\n\n\n<li>Provide no SEO benefit<\/li>\n<\/ul>\n\n\n\n<p>If your analytics show unexplained bandwidth surges or CPU spikes without a rise in human traffic, AI crawlers may be the culprit.<\/p>\n\n\n\n<p>These bots won\u2019t disappear anytime soon, so adapting your robots.txt strategy now will save headaches later. The next section covers exactly how to do that.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Blocking AI Crawlers with Robots.txt and LLMS.txt<\/h2>\n\n\n\n<p>When performance and data ownership are at stake, blocking AI crawlers can make an immediate difference. The good news is that many reputable AI providers follow robots.txt directives. Extending your existing rules is a simple way to control this traffic.<\/p>\n\n\n\n<p>Most major AI providers, including OpenAI and Anthropic, respect <code>robots.txt<\/code> rules. To block their crawlers, add specific disallow directives:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono\" style=\"font-size:.875rem;font-family:Code-Pro-JetBrains-Mono,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:1.25rem;--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: GPTBot\nDisallow: \/\n\nUser-agent: ClaudeBot\nDisallow: \/<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2m-6 9l2 2 4-4\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">GPTBot<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/<\/span><\/span>\n<span class=\"line\"><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">ClaudeBot<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>For emerging AI bots, check their published user-agent strings. You can also maintain a separate list for quick updates.<\/p>\n\n\n\n<p>A newer complementary standard\u2014<code>llms.txt<\/code>\u2014is gaining traction. While not officially standardized, it allows site owners to express whether their content can be used for AI training.<\/p>\n\n\n\n<p>Example:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono\" style=\"font-size:.875rem;font-family:Code-Pro-JetBrains-Mono,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:1.25rem;--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: GPTBot\nDisallow: \/\nAI-Policy: disallow<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2m-6 9l2 2 4-4\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">GPTBot<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">AI-Policy:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">disallow<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>Together, these tools let you maintain visibility in search engines while protecting your site from excessive AI scraping.<\/p>\n\n\n\n<p>With your crawler rules in place, the next level of protection comes from your hosting environment itself, starting with server-level automation.<\/p>\n\n\n\n<figure id=\"Feature-compare-robots.txt-to-llms.txt\" class=\"wp-block-table is-style-stripes\"><table><thead><tr><th class=\"has-text-align-left\" data-align=\"left\"><strong>Feature<\/strong><\/th><th class=\"has-text-align-left\" data-align=\"left\"><strong>robots.txt<\/strong><\/th><th class=\"has-text-align-left\" data-align=\"left\"><strong>llms.txt<\/strong><\/th><\/tr><\/thead><tbody><tr><td class=\"has-text-align-left\" data-align=\"left\"><strong>Purpose<\/strong><\/td><td class=\"has-text-align-left\" data-align=\"left\">Tells search engines what they can crawl.<\/td><td class=\"has-text-align-left\" data-align=\"left\">Tells AI crawlers how they can use your content.<\/td><\/tr><tr><td class=\"has-text-align-left\" data-align=\"left\"><strong>Used By<\/strong><\/td><td class=\"has-text-align-left\" data-align=\"left\">Googlebot, Bingbot, and other search bots.<\/td><td class=\"has-text-align-left\" data-align=\"left\">GPTBot, ClaudeBot, and other AI crawlers.<\/td><\/tr><tr><td class=\"has-text-align-left\" data-align=\"left\"><strong>Standard<\/strong><\/td><td class=\"has-text-align-left\" data-align=\"left\">Long-established and widely supported.<\/td><td class=\"has-text-align-left\" data-align=\"left\">New and still developing.<\/td><\/tr><tr><td class=\"has-text-align-left\" data-align=\"left\"><strong>Location<\/strong><\/td><td class=\"has-text-align-left\" data-align=\"left\"><code>yourdomain.com\/robots.txt<\/code><\/td><td class=\"has-text-align-left\" data-align=\"left\"><code>yourdomain.com\/llms.txt<\/code><\/td><\/tr><tr><td class=\"has-text-align-left\" data-align=\"left\"><strong>Affects SEO<\/strong><\/td><td class=\"has-text-align-left\" data-align=\"left\">Yes, it controls what gets indexed.<\/td><td class=\"has-text-align-left\" data-align=\"left\">No, it focuses on AI data use, not rankings.<\/td><\/tr><tr><td class=\"has-text-align-left\" data-align=\"left\"><strong>Main Benefit<\/strong><\/td><td class=\"has-text-align-left\" data-align=\"left\">Saves crawler allocation and improves performance.<\/td><td class=\"has-text-align-left\" data-align=\"left\">Reduces unwanted AI scraping.<\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">Advanced Control: Combining Robots.txt with Server Rules<\/h2>\n\n\n\n<p>Even the best robots.txt file can\u2019t stop bots that ignore it completely. For these rogue crawlers, the solution lives at the server level. Combining robots.txt with .htaccess rules gives you enforcement power rather than polite requests.<\/p>\n\n\n\n<p>While <code>robots.txt<\/code> works for compliant crawlers, rogue bots often ignore it. For those situations, you need stronger enforcement at the server level. Your <code>.htaccess<\/code> file (a configuration file that controls how your web server handles requests) can physically block unwanted bots before they ever reach your site.<\/p>\n\n\n\n<p>Example:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono\" style=\"font-size:.875rem;font-family:Code-Pro-JetBrains-Mono,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:1.25rem;--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>&lt;IfModule mod_rewrite.c&gt;\nRewriteEngine On\nRewriteCond %{HTTP_USER_AGENT} (AhrefsBot|MJ12bot|Baiduspider) [NC]\nRewriteRule .* - [F,L]\n&lt;\/IfModule&gt;<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2m-6 9l2 2 4-4\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M9 5H7a2 2 0 00-2 2v12a2 2 0 002 2h10a2 2 0 002-2V7a2 2 0 00-2-2h-2M9 5a2 2 0 002 2h2a2 2 0 002-2M9 5a2 2 0 012-2h2a2 2 0 012 2\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #E1E4E8\">&lt;<\/span><span style=\"color: #B392F0\">IfModule<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">mod_rewrite.c<\/span><span style=\"color: #E1E4E8\">&gt;<\/span><\/span>\n<span class=\"line\"><span style=\"color: #F97583\">RewriteEngine<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #B392F0\">On<\/span><\/span>\n<span class=\"line\"><span style=\"color: #F97583\">RewriteCond<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #DBEDFF\">%{HTTP_USER_AGENT}<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">(AhrefsBot|MJ12bot|Baiduspider)<\/span><span style=\"color: #E1E4E8\"> [NC]<\/span><\/span>\n<span class=\"line\"><span style=\"color: #F97583\">RewriteRule<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #DBEDFF\">.*<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">-<\/span><span style=\"color: #E1E4E8\"> [F,L]<\/span><\/span>\n<span class=\"line\"><span style=\"color: #E1E4E8\">&lt;\/<\/span><span style=\"color: #B392F0\">IfModule<\/span><span style=\"color: #E1E4E8\">&gt;<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>This blocks unwanted user agents at the server level before they consume resources.<\/p>\n\n\n\n<p>Server-level rules require careful testing. If you\u2019re not comfortable editing <code>.htaccess<\/code> directly, contact your hosting provider\u2019s support team. Your hosting support experts can help implement these rules safely, ensuring you block the right bots without affecting legitimate traffic.<\/p>\n\n\n\n<p>InMotion Hosting\u2019s infrastructure supports both .htaccess rules and firewall configurations, letting you manage good and bad traffic efficiently without downtime.<\/p>\n\n\n\n<p>Once your defenses are set, it\u2019s helpful to see how these techniques work in practice. The following section explores a few real-world scenarios that illustrate the balance between visibility and protection.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Real-World Use Cases<\/h2>\n\n\n\n<p>Theory only goes so far, seeing how these rules apply in real scenarios brings them to life. Whether you\u2019re running an e-commerce store or preparing for a redesign, knowing how to control bots in context makes your strategy stronger.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">1. Website Redesign<\/h3>\n\n\n\n<p>Before launching a redesign, use robots.txt to block crawlers from the staging version:<\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono-NL.ttf\" style=\"font-size:clamp(16px, 1rem, 24px);font-family:Code-Pro-JetBrains-Mono-NL,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:clamp(26px, 1.625rem, 39px);--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: *\nDisallow: \/<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M4.5 12.75l6 6 9-13.5\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M16.5 8.25V6a2.25 2.25 0 00-2.25-2.25H6A2.25 2.25 0 003.75 6v8.25A2.25 2.25 0 006 16.5h2.25m8.25-8.25H18a2.25 2.25 0 012.25 2.25V18A2.25 2.25 0 0118 20.25h-7.5A2.25 2.25 0 018.25 18v-1.5m8.25-8.25h-6a2.25 2.25 0 00-2.25 2.25v6\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #79B8FF\">*<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<p>Then remove the rule immediately after launch to restore indexing.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">2. E-Commerce Filtering Pages<\/h3>\n\n\n\n<p>Retail sites often generate duplicate URLs from filtering options (e.g., \/shirts?color=blue). Use robots.txt to prevent crawlers from indexing duplicate variations.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">3. Content Migration<\/h3>\n\n\n\n<p>During content audits or CMS transitions, temporarily block old directories to avoid index bloat, then reopen access once redirects are live.<\/p>\n\n\n\n<p>Each example helps maintain SEO integrity while keeping crawl efficiency focused on high-value pages.<\/p>\n\n\n\n<p>Now that we\u2019ve seen these examples in action, let\u2019s cover the mistakes site owners make most often and how to avoid them.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Common Mistakes With Robots.txt<\/h2>\n\n\n\n<p>Every webmaster makes at least one crawl control mistake. Most are easy to fix but can be costly if left unchecked. Being aware of these pitfalls early helps you catch issues before they affect traffic or visibility.<\/p>\n\n\n\n<p><strong>Accidentally blocking your entire site:<\/strong><\/p>\n\n\n\n<div class=\"wp-block-kevinbatdorf-code-block-pro\" data-code-block-pro-font-family=\"Code-Pro-JetBrains-Mono-NL.ttf\" style=\"font-size:clamp(16px, 1rem, 24px);font-family:Code-Pro-JetBrains-Mono-NL,ui-monospace,SFMono-Regular,Menlo,Monaco,Consolas,monospace;line-height:clamp(26px, 1.625rem, 39px);--cbp-tab-width:2;tab-size:var(--cbp-tab-width, 2)\"><span role=\"button\" tabindex=\"0\" style=\"color:#e1e4e8;display:none\" aria-label=\"Copy\" class=\"code-block-pro-copy-button\"><pre class=\"code-block-pro-copy-button-pre\" aria-hidden=\"true\"><textarea class=\"code-block-pro-copy-button-textarea\" tabindex=\"-1\" aria-hidden=\"true\" readonly>User-agent: *\nDisallow: \/<\/textarea><\/pre><svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" style=\"width:24px;height:24px\" fill=\"none\" viewbox=\"0 0 24 24\" stroke=\"currentColor\" stroke-width=\"2\"><path class=\"with-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M4.5 12.75l6 6 9-13.5\"><\/path><path class=\"without-check\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"M16.5 8.25V6a2.25 2.25 0 00-2.25-2.25H6A2.25 2.25 0 003.75 6v8.25A2.25 2.25 0 006 16.5h2.25m8.25-8.25H18a2.25 2.25 0 012.25 2.25V18A2.25 2.25 0 0118 20.25h-7.5A2.25 2.25 0 018.25 18v-1.5m8.25-8.25h-6a2.25 2.25 0 00-2.25 2.25v6\"><\/path><\/svg><\/span><pre class=\"shiki github-dark\" style=\"background-color: #24292e\" tabindex=\"0\"><code><span class=\"line\"><span style=\"color: #B392F0\">User-agent:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #79B8FF\">*<\/span><\/span>\n<span class=\"line\"><span style=\"color: #B392F0\">Disallow:<\/span><span style=\"color: #E1E4E8\"> <\/span><span style=\"color: #9ECBFF\">\/<\/span><\/span><\/code><\/pre><\/div>\n\n\n\n<ul class=\"wp-block-list\">\n<li>\u00a0Always verify before publishing.<\/li>\n\n\n\n<li><strong>Using robots.txt for privacy:<\/strong> Anyone can view it at \/robots.txt. Never rely on it to hide sensitive data.<\/li>\n\n\n\n<li><strong>Leaving staging rules in production:<\/strong> Double-check after launches or migrations.<\/li>\n\n\n\n<li><strong>Blocking assets:<\/strong> Ensure your site\u2019s design and JavaScript remain accessible to Googlebot for accurate rendering.<\/li>\n<\/ul>\n\n\n\n<p>Avoiding these common traps keeps your crawl rules clean and effective. Once you have a stable configuration, the next step is ensuring your hosting setup can handle the traffic that does get through.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Performance and SEO Implications<\/h2>\n\n\n\n<p>Good crawl control doesn\u2019t just organize bots, it protects your performance metrics. Every second saved on server response time contributes to better user experience and stronger search visibility.<\/p>\n\n\n\n<p>By combining robots.txt control with performance-optimized infrastructure (like NVMe-powered servers and dedicated hosting) you protect both user experience and SEO outcomes.<\/p>\n\n\n\n<p>We found that websites on dedicated environments <a href=\"https:\/\/www.inmotionhosting.com\/blog\/ai-crawlers-slowing-down-your-website\/\">recover up to 60% faster<\/a> from crawler-related slowdowns compared to shared hosting setups.<\/p>\n\n\n\n<p>Performance and SEO go hand in hand. When you reduce bot overhead, you improve everything else that matters to your users. Let\u2019s close with a few habits that keep your crawl strategy strong over time.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/www.inmotionhosting.com\/blog\/wp-content\/uploads\/2025\/10\/Blue-and-White-Simple-Daily-Vlogger-YouTube-Banner-3-1024x576.png\" alt=\"Best Practices for Crawl Management\" class=\"wp-image-81271\"><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">Best Practices for Ongoing Crawl Management<\/h2>\n\n\n\n<p>Long-term success with robots.txt depends on maintenance. As new bots and technologies appear, your file should evolve to reflect them. Treat it like any other system configuration. You\u2019ll need to review, test, and update regularly.<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Review robots.txt quarterly or after structural changes.<\/li>\n\n\n\n<li>Log and monitor unusual crawler activity in analytics or server logs.<\/li>\n\n\n\n<li>Use Google Search Console\u2019s Crawl Stats report to track patterns.<\/li>\n\n\n\n<li>Pair robots.txt with caching and CDN layers for optimal speed.<\/li>\n\n\n\n<li>Stay current on new user agents and update disallow lists as needed.<\/li>\n<\/ol>\n\n\n\n<p>Crawl control isn\u2019t a one-time fix, it\u2019s an ongoing practice that supports the health of your entire site. When paired with reliable hosting and regular reviews, it keeps performance stable and SEO strong.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>Last Thoughts<\/strong><\/h2>\n\n\n\n<p>robots.txt gives you the power to direct how bots interact with your website. This includes improving SEO focus, conserving server resources, and protecting performance. By using it strategically alongside tools like .htaccess and emerging standards such as llms.txt, you can maintain visibility where it matters and safeguard your hosting environment from wasteful traffic.<\/p>\n\n\n\n<p>Managing crawl access is part of modern website hygiene. It balances discovery with control, helping your business stay fast, visible, and secure as the web evolves. With thoughtful configuration and the right infrastructure behind you, your site will always be ready for whatever bots come next.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Crawlers drive search visibility, but they can also overload your server if left unchecked. This guide shows you how to use robots.txt to take control. Learn how to block low-value directories, protect essential assets that Google needs to render your site, manage crawl rates with delays, and defend against <a href=\"https:\/\/www.inmotionhosting.com\/blog\/ai-crawlers-slowing-down-your-website\/\">bandwidth-heavy AI bots<\/a>. You&#8217;ll get practical examples, real-world use cases, and best practices for combining robots.txt with server rules. Strategic crawler management improves site speed, reduces hosting costs, and ensures search engines focus on your most important content.<\/p>\n","protected":false},"author":57037,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[4288],"tags":[],"class_list":["post-130930","post","type-post","status-publish","format-standard","hentry","category-website"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.1.1 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>How To Use Robots.txt | InMotion Hosting<\/title>\n<meta name=\"description\" content=\"Control bots with robots.txt. Block unwanted crawlers, protect site speed, and keep your SEO performance strong.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"How To Use Robots.txt | InMotion Hosting\" \/>\n<meta property=\"og:description\" content=\"Control bots with robots.txt. Block unwanted crawlers, protect site speed, and keep your SEO performance strong.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\" \/>\n<meta property=\"og:site_name\" content=\"InMotion Hosting Support Center\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/inmotionhosting\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-10-16T12:55:16+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2025-11-12T22:28:33+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1200\" \/>\n\t<meta property=\"og:image:height\" content=\"630\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Sam Page\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@InMotionHosting\" \/>\n<meta name=\"twitter:site\" content=\"@InMotionHosting\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Sam Page\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"12 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\"},\"author\":{\"name\":\"Sam Page\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/person\/18aba2647808ba3ddaf4406008b92ec1\"},\"headline\":\"How to Use robots.txt Disallow to Block Crawlers and Protect Site Performance\",\"datePublished\":\"2025-10-16T12:55:16+00:00\",\"dateModified\":\"2025-11-12T22:28:33+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\"},\"wordCount\":2590,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#organization\"},\"image\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-1024x538.png\",\"articleSection\":[\"Website\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\",\"url\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\",\"name\":\"How To Use Robots.txt | InMotion Hosting\",\"isPartOf\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-1024x538.png\",\"datePublished\":\"2025-10-16T12:55:16+00:00\",\"dateModified\":\"2025-11-12T22:28:33+00:00\",\"description\":\"Control bots with robots.txt. Block unwanted crawlers, protect site speed, and keep your SEO performance strong.\",\"breadcrumb\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage\",\"url\":\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png\",\"contentUrl\":\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png\",\"width\":1200,\"height\":630},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/www.inmotionhosting.com\/support\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"How to Use robots.txt Disallow to Block Crawlers and Protect Site Performance\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#website\",\"url\":\"https:\/\/www.inmotionhosting.com\/support\/\",\"name\":\"InMotion Hosting Support Center\",\"description\":\"Web Hosting Support &amp; Tutorials\",\"publisher\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.inmotionhosting.com\/support\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#organization\",\"name\":\"InMotion Hosting\",\"url\":\"https:\/\/www.inmotionhosting.com\/support\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2023\/02\/inmotion-hosting-logo-yoast.jpg\",\"contentUrl\":\"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2023\/02\/inmotion-hosting-logo-yoast.jpg\",\"width\":696,\"height\":696,\"caption\":\"InMotion Hosting\"},\"image\":{\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/inmotionhosting\/\",\"https:\/\/x.com\/InMotionHosting\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/person\/18aba2647808ba3ddaf4406008b92ec1\",\"name\":\"Sam Page\",\"url\":\"https:\/\/www.inmotionhosting.com\/support\/author\/samp\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"How To Use Robots.txt | InMotion Hosting","description":"Control bots with robots.txt. Block unwanted crawlers, protect site speed, and keep your SEO performance strong.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/","og_locale":"en_US","og_type":"article","og_title":"How To Use Robots.txt | InMotion Hosting","og_description":"Control bots with robots.txt. Block unwanted crawlers, protect site speed, and keep your SEO performance strong.","og_url":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/","og_site_name":"InMotion Hosting Support Center","article_publisher":"https:\/\/www.facebook.com\/inmotionhosting\/","article_published_time":"2025-10-16T12:55:16+00:00","article_modified_time":"2025-11-12T22:28:33+00:00","og_image":[{"width":1200,"height":630,"url":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png","type":"image\/png"}],"author":"Sam Page","twitter_card":"summary_large_image","twitter_creator":"@InMotionHosting","twitter_site":"@InMotionHosting","twitter_misc":{"Written by":"Sam Page","Est. reading time":"12 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#article","isPartOf":{"@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/"},"author":{"name":"Sam Page","@id":"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/person\/18aba2647808ba3ddaf4406008b92ec1"},"headline":"How to Use robots.txt Disallow to Block Crawlers and Protect Site Performance","datePublished":"2025-10-16T12:55:16+00:00","dateModified":"2025-11-12T22:28:33+00:00","mainEntityOfPage":{"@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/"},"wordCount":2590,"commentCount":0,"publisher":{"@id":"https:\/\/www.inmotionhosting.com\/support\/#organization"},"image":{"@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage"},"thumbnailUrl":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-1024x538.png","articleSection":["Website"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/","url":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/","name":"How To Use Robots.txt | InMotion Hosting","isPartOf":{"@id":"https:\/\/www.inmotionhosting.com\/support\/#website"},"primaryImageOfPage":{"@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage"},"image":{"@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage"},"thumbnailUrl":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41-1024x538.png","datePublished":"2025-10-16T12:55:16+00:00","dateModified":"2025-11-12T22:28:33+00:00","description":"Control bots with robots.txt. Block unwanted crawlers, protect site speed, and keep your SEO performance strong.","breadcrumb":{"@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#primaryimage","url":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png","contentUrl":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2025\/10\/Technology-How-To-Solid-3-Jacksons-Purple-41.png","width":1200,"height":630},{"@type":"BreadcrumbList","@id":"https:\/\/www.inmotionhosting.com\/support\/website\/how-to-use-robots-txt-to-block-crawlers\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/www.inmotionhosting.com\/support\/"},{"@type":"ListItem","position":2,"name":"How to Use robots.txt Disallow to Block Crawlers and Protect Site Performance"}]},{"@type":"WebSite","@id":"https:\/\/www.inmotionhosting.com\/support\/#website","url":"https:\/\/www.inmotionhosting.com\/support\/","name":"InMotion Hosting Support Center","description":"Web Hosting Support &amp; Tutorials","publisher":{"@id":"https:\/\/www.inmotionhosting.com\/support\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.inmotionhosting.com\/support\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/www.inmotionhosting.com\/support\/#organization","name":"InMotion Hosting","url":"https:\/\/www.inmotionhosting.com\/support\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/logo\/image\/","url":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2023\/02\/inmotion-hosting-logo-yoast.jpg","contentUrl":"https:\/\/www.inmotionhosting.com\/support\/wp-content\/uploads\/2023\/02\/inmotion-hosting-logo-yoast.jpg","width":696,"height":696,"caption":"InMotion Hosting"},"image":{"@id":"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/inmotionhosting\/","https:\/\/x.com\/InMotionHosting"]},{"@type":"Person","@id":"https:\/\/www.inmotionhosting.com\/support\/#\/schema\/person\/18aba2647808ba3ddaf4406008b92ec1","name":"Sam Page","url":"https:\/\/www.inmotionhosting.com\/support\/author\/samp\/"}]}},"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"primary_category":{"id":4288,"name":"Website","slug":"website","link":"https:\/\/www.inmotionhosting.com\/support\/website\/"},"_links":{"self":[{"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/posts\/130930","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/users\/57037"}],"replies":[{"embeddable":true,"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/comments?post=130930"}],"version-history":[{"count":9,"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/posts\/130930\/revisions"}],"predecessor-version":[{"id":131121,"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/posts\/130930\/revisions\/131121"}],"wp:attachment":[{"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/media?parent=130930"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/categories?post=130930"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.inmotionhosting.com\/support\/wp-json\/wp\/v2\/tags?post=130930"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}