UNC-Libraries/hy-c

View on GitHub
public/robots.txt

Summary

Maintainability
Test Coverage
# See http://www.robotstxt.org/robotstxt.html for documentation on how to use the robots.txt file
#
# To ban all spiders from the entire site uncomment the next two lines:
# User-agent: *
# Disallow: /
User-agent: *
Crawl-delay: 8
Disallow: /catalog
Disallow: /files/*/stats
Disallow: /works/*/stats

# allow google scholar to crawl the site as per https://scholar.google.com/intl/en/scholar/inclusion.html#crawl
User-agent: Googlebot
Allow: /

# resource sync sitemap
Sitemap: https://cdr.lib.unc.edu/resourcelist.xml