# Robots.txt for crawler User-agent: * # Disallow Crawler Disallow: /User Disallow: /Dateien Disallow: /Nachrichten/Suche Disallow: /*?bPrint=true* Disallow: /Nachrichten/Archiv-Liste_*.html Disallow: /*/Zusammengefasst-*.html Disallow: /Zusammengefasst-*.html # Crawler often creates invalid script/webresource resource request Disallow: /ScriptResource Disallow: /WebResource User-agent: backlink-check.de Disallow: / User-agent: BacklinkCrawler Disallow: / User-agent: ExtractorPro Disallow: / User-agent: Fasterfox Disallow: / User-agent: LinkextractorPro Disallow: / User-agent: LinkWalker Disallow: / User-agent: MJ12bot Disallow: / User-agent: Openbot Disallow: / User-agent: rogerbot Disallow: / User-agent: searchpreview Disallow: / User-agent: SemrushBot Disallow: / User-agent: SEODAT Disallow: / User-agent: SEOENGBot Disallow: / User-agent: SEOkicks-Robot Disallow: / User-agent: True_Robot Disallow: / User-agent: URL Control Disallow: / User-agent: URL_Spider_Pro Disallow: / User-agent: xovi Disallow: / User-agent: GPTBot Disallow: / User-agent: ChatGPT-User Disallow: / User-agent: CCBot Disallow: / # Uber Metrics User-agent: um-IC Disallow: / User-agent: Google-Extended Disallow: / # Max crawler Time per page in sec Crawl-Delay: 2 #Sitemap Sitemap: https://www.nord24.de/Sitemap_Index.xml.gz