Sitemap: https://www.nuroa.com.mx/sitemap/sitemap_index.xml # will this work? User-agent: * Disallow: /adform/ Disallow: /listing/ Disallow: /detail/ Disallow: /project/ Disallow: /publisher/ Disallow: /controller Disallow: /controller/ Disallow: /ajax Disallow: /ajax/ Disallow: /conversion Disallow: /conversion/ Disallow: /dynamic Disallow: /dynamic/ Disallow: /*/ajax Disallow: /*/ajax/ Disallow: /admin Disallow: /admin/ Disallow: /*/admin Disallow: /*/admin/ Disallow: /dashboard Disallow: /dashboard/ Disallow: /*/dashboard Disallow: /*/dashboard/ Disallow: /booking Disallow: /booking/ Disallow: /*/booking Disallow: /*/booking/ Disallow: /review Disallow: /review/ Disallow: /*/review Disallow: /*/review/ Disallow: /ad/ Disallow: /*/ad/ Disallow: /property/ Disallow: /*/property/ Disallow: /details/ Disallow: /*/details/ Disallow: /ad-list-* Disallow: /*/ad-list-* Disallow: /rss Disallow: /*/rss Disallow: /api/ # Crawlers that are kind enough to obey, but which we'd rather not have # unless they're feeding search engines. User-agent: UbiCrawler Disallow: / User-agent: DOC Disallow: / User-agent: Zao Disallow: / User-agent: Jyxobot Disallow: / User-agent: SISTRIX Crawler Disallow: / # Some bots are known to be trouble, particularly those designed to copy # entire sites. Please obey robots.txt. User-agent: sitecheck.internetseer.com Disallow: / User-agent: Zealbot Disallow: / User-agent: MSIECrawler Disallow: / User-agent: SiteSnagger Disallow: / User-agent: WebStripper Disallow: / User-agent: WebCopier Disallow: / User-agent: Fetch Disallow: / User-agent: Offline Explorer Disallow: / User-agent: Teleport Disallow: / User-agent: TeleportPro Disallow: / User-agent: WebZIP Disallow: / User-agent: linko Disallow: / User-agent: HTTrack Disallow: / User-agent: Microsoft.URL.Control Disallow: / User-agent: Xenu Disallow: / User-agent: larbin Disallow: / User-agent: libwww Disallow: / User-agent: ZyBORG Disallow: / User-agent: Download Ninja Disallow: / # # Sorry, wget in its recursive mode is a frequent problem. # Please read the man page and use it properly; there is a # --wait option you can use to set the delay between hits, # for instance. # User-agent: wget Disallow: / # # The 'grub' distributed client has been *very* poorly behaved. # User-agent: grub-client Disallow: / # # Doesn't follow robots.txt anyway, but... # User-agent: k2spider Disallow: / # # Hits many times per second, not acceptable # http://www.nameprotect.com/botinfo.html User-agent: NPBot Disallow: / # A capture bot, downloads gazillions of pages with no public benefit # http://www.webreaper.net/ User-agent: WebReaper Disallow: /
Folgende Keywords wurden erkannt. Überprüfe die Optimierung dieser Keywords für Deine Seite.
(Nice to have)