# https://www.robotstxt.org/robotstxt.html User-agent: * Allow: / # Block sensitive and unnecessary routes Disallow: /api/ Disallow: /functions/ Disallow: /_next/ Disallow: /_error Disallow: /admin/ Disallow: /login/ Disallow: /signup/ Disallow: /settings/ # Block search query pages to prevent duplicate content issues Disallow: /search # Block non-canonical versions of dynamic routes Disallow: /categories/*?* Disallow: /products/*?* # Sitemap location Sitemap: https://litreviews.co/sitemap.xml # Security: Block crawlers from accessing private data Disallow: /private/ Disallow: /tmp/ Disallow: /hidden/ # Security: Prevent indexing of unnecessary assets Disallow: /*.js$ Disallow: /*.css$ Disallow: /*.map$ Disallow: /*.json$ # Prevent specific bots User-agent: SemrushBot Disallow: / User-agent: AhrefsBot Disallow: / User-agent: MJ12bot Disallow: / # Crawl-delay for all bots Crawl-delay: 2 # Specific directives for known bots User-agent: Googlebot Crawl-delay: 1 Allow: / User-agent: Bingbot Crawl-delay: 1 Allow: / User-agent: Slurp Crawl-delay: 1 Allow: / # AI and bot-specific rules User-agent: AIbot Disallow: /sensitive/ # Advanced SEO techniques User-agent: Mediapartners-Google Allow: / # Ensuring no index for certain pages Disallow: /noindex/ # Honeypot for security Disallow: /secure/logins.html # Allow specific important pages Allow: /categories/ Allow: /products/ Allow: /latest/ Allow: /about/ Allow: /contact/ Allow: /privacy/ # Allow product images to be crawled Allow: /products/*.png$ Allow: /products/*.jpg$ Allow: /products/*.jpeg$ Allow: /products/*.webp$