User-agent: * # Disallow: /query/ # Disallow: /noindex/ Disallow: /empresa/index/ # Disallow: /empresa/oferta/ # Disallow: /empresa/mensaje/ # Disallow: /empresa/candidato/ Disallow: /admin/ Disallow: /candidato/ Disallow: /phpmyadmin/ Disallow: /cgi-bin/ Sitemap: https://hosteleo.com/sitemap.xml Sitemap: https://hosteleo.com/sitemap-categorias.xml Sitemap: https://hosteleo.com/sitemap-tags.xml Sitemap: https://hosteleo.com/sitemap-empresas.xml Sitemap: https://hosteleo.com/sitemap-provincias.xml # # A partir de aquĆ­ es opcional pero recomendado. # # Lista de bots que suelen respetar el robots.txt pero rara # vez hacen un buen uso del sitio y abusan bastante... # Anyadir al gusto del consumidor... User-agent: MSIECrawler Disallow: / User-agent: WebCopier Disallow: / User-agent: HTTrack Disallow: / User-agent: Microsoft.URL.Control Disallow: / User-agent: libwww Disallow: / User-agent: 80legs Disallow: / User-agent: Ezooms Disallow: / User-agent: Baiduspider Disallow: / User-agent: AhrefsBot Disallow: / User-agent: YandexImages Disallow: / User-agent: YandexBot Disallow: / User-agent: Sogou Disallow: / User-agent: MJ12bot Disallow: / User-agent: SeznamBot Disallow: / User-agent: WBSearchBot Disallow: / User-agent: Exabot Disallow: / User-agent: SISTRIX Disallow: / User-agent: JikeSpider Disallow: / User-agent: Sosospider Disallow: / User-agent: 360Spider Disallow: / User-agent: proximic Disallow: / User-agent: Orthogaffe Disallow: / User-agent: UbiCrawler Disallow: / User-agent: sitecheck.internetseer.com Disallow: / User-agent: Zealbot Disallow: / User-agent: SiteSnagger Disallow: / User-agent: WebStripper Disallow: / User-agent: WebCopier Disallow: / User-agent: Fetch Disallow: / User-agent: Offline Explorer Disallow: / User-agent: Teleport Disallow: / User-agent: TeleportPro Disallow: / User-agent: WebZIP Disallow: / User-agent: linko Disallow: / User-agent: HTTrack Disallow: / User-agent: Xenu Disallow: / User-agent: larbin Disallow: / User-agent: ZyBORG Disallow: / User-agent: Download Ninja Disallow: / # Crawlers that are kind enough to obey, but which we'd rather not have # unless they're feeding search engines. User-agent: UbiCrawler Disallow: / User-agent: DOC Disallow: / User-agent: Zao Disallow: / # Sorry, wget in its recursive mode is a frequent problem. # Please read the man page and use it properly; there is a # --wait option you can use to set the delay between hits, # for instance. # User-agent: wget Disallow: / # # The 'grub' distributed client has been *very* poorly behaved. # User-agent: grub-client Disallow: / # # Doesn't follow robots.txt anyway, but... # User-agent: k2spider Disallow: / # # Hits many times per second, not acceptable # http://www.nameprotect.com/botinfo.html User-agent: NPBot Disallow: / # A capture bot, downloads gazillions of pages with no public benefit # http://www.webreaper.net/ User-agent: WebReaper Disallow: / User-agent: noxtrumbot Crawl-delay: 50 User-agent: msnbot Crawl-delay: 30 User-agent: Slurp Crawl-delay: 10 User-agent: Mediapartners-Google Disallow: