# Block known resource-heavy and AI data-scraping bots User-agent: MJ12bot Disallow: / User-agent: AhrefsBot Disallow: / User-agent: PetalBot Disallow: / User-agent: BLEXBot Disallow: / User-agent: DotBot Disallow: / User-agent: SemrushBot Disallow: / User-agent: Sogou Disallow: / User-agent: Baiduspider Disallow: / User-agent: Yandex Disallow: / User-agent: MegaIndex.ru Disallow: / User-agent: MauiBot Disallow: / User-agent: GPTBot Disallow: / User-agent: GPTBot/1.0 Disallow: / User-agent: CCBot Disallow: / User-agent: ia_archiver Disallow: / User-agent: Screaming Frog SEO Spider Disallow: / User-agent: ZoominfoBot Disallow: / User-agent: Exabot Disallow: / User-agent: ByteSpider Disallow: / User-agent: ClaudeBot Disallow: / User-agent: PerplexityBot Disallow: / # Block generic scrapers and request libraries User-agent: python-requests Disallow: / User-agent: curl Disallow: / User-agent: wget Disallow: / User-agent: libwww Disallow: / User-agent: Java Disallow: / User-agent: Go-http-client Disallow: / # Default restrictions for all other crawlers User-agent: * Crawl-delay: 5 Disallow: /wp-admin/ Disallow: /wp-login.php Disallow: /xmlrpc.php Disallow: /cgi-bin/ Disallow: /trackback/ Disallow: /comments/feed/ Disallow: /?s= Disallow: /*?s=* Allow: /wp-admin/admin-ajax.php # Help crawlers locate your sitemap Sitemap: https://yourdomain.com/sitemap.xml