# https://www.robotstxt.org/robotstxt.html
# AllDevTool.com - Advanced Robots.txt for SEO Optimization
# Default rules for all crawlers
User-agent: *
Allow: /
# Crawl delay for polite crawling (1 second)
Crawl-delay: 1
# Allow important directories for SEO
Allow: /tools/
Allow: /discover/
Allow: /category/
Allow: /collections/
Allow: /guides/
Allow: /api/sitemap
Allow: /_next/static/
Allow: /images/
Allow: /screenshots/
# Prevent crawling of redirected URLs (these redirect to final URLs)
Disallow: /tools/html-formatter
Disallow: /tools/yaml-to-json
Disallow: /tools/csv-to-json
Disallow: /tools/json-to-csv
Disallow: /tools/xml-to-json
# Don't crawl non-existent pages
Disallow: /tools/websocket-tester
# Disallow paths that should not be indexed
Disallow: /api/
Disallow: /admin/
Disallow: /private/
Disallow: /_next/
Disallow: /ads-debug/
Disallow: /ad-test/
Disallow: /temp/
Disallow: /draft/
Disallow: /test/
Disallow: /*.json$
Disallow: /*?print=*
Disallow: /*?debug=*
Disallow: /*&print=*
Disallow: /*&debug=*
# Block specific file types that shouldn't be indexed
Disallow: /*.pdf$
Disallow: /*.doc$
Disallow: /*.docx$
Disallow: /*.xls$
Disallow: /*.xlsx$
Disallow: /*.ppt$
Disallow: /*.pptx$
# Specific instructions for Googlebot (most important)
User-agent: Googlebot
Allow: /
Crawl-delay: 1
# Allow Google to index images
Allow: /images/
Allow: /screenshots/
Allow: /_next/static/images/
# Specific instructions for Bingbot
User-agent: Bingbot
Allow: /
Crawl-delay: 1
# Specific instructions for other major search engines
User-agent: Slurp
Allow: /
Crawl-delay: 2
User-agent: DuckDuckBot
Allow: /
Crawl-delay: 1
User-agent: Baiduspider
Allow: /
Crawl-delay: 2
User-agent: YandexBot
Allow: /
Crawl-delay: 1
# Allow social media crawlers for better sharing
User-agent: facebookexternalhit
Allow: /
User-agent: Twitterbot
Allow: /
User-agent: LinkedInBot
Allow: /
User-agent: WhatsApp
Allow: /
User-agent: TelegramBot
Allow: /
# Allow SEO and analytics tools
User-agent: AhrefsBot
Allow: /
Crawl-delay: 1
User-agent: SemrushBot
Allow: /
Crawl-delay: 2
User-agent: MJ12bot
Allow: /
Crawl-delay: 3
User-agent: DotBot
Allow: /
Crawl-delay: 2
# Allow Google's various specialized bots
User-agent: Googlebot-Image
Allow: /images/
Allow: /screenshots/
Allow: /_next/static/images/
User-agent: Googlebot-Video
Allow: /
User-agent: Google-InspectionTool
Allow: /
User-agent: GoogleOther
Allow: /
# Block aggressive crawlers that can harm performance
User-agent: BLEXBot
Disallow: /
User-agent: DataForSeoBot
Disallow: /
User-agent: PetalBot
Disallow: /
User-agent: CCBot
Disallow: /
User-agent: GPTBot
Disallow: /
User-agent: ChatGPT-User
Disallow: /
User-agent: anthropic-ai
Disallow: /
User-agent: Claude-Web
Disallow: /
# Block malicious bots and scrapers
User-agent: SiteAuditBot
Disallow: /
User-agent: AhrefsSiteAudit
Disallow: /
User-agent: spbot
Disallow: /
User-agent: MegaIndex
Disallow: /
User-agent: dotbot
Disallow: /
# Sitemaps (multiple formats for comprehensive coverage)
Sitemap: https://www.alldevtool.com/sitemap.xml
Sitemap: https://www.alldevtool.com/api/sitemap
# HTML sitemap for crawlers
Sitemap: https://www.alldevtool.com/sitemap-page
# Text sitemap with all URLs listed
Sitemap: https://www.alldevtool.com/sitemap-text.txt
# Static HTML index with all tools for emergency crawling
Sitemap: https://www.alldevtool.com/tool-index.html
# Additional sitemap for images (if implemented)
# Sitemap: https://www.alldevtool.com/image-sitemap.xml
# Host directive (helps with canonical domain)
Host: https://www.alldevtool.com