robots.txt parser
HTTP /api/v1/http/robotsFetch and parse robots.txt — User-agent groups, Disallow/Allow, Sitemap, Crawl-delay.
https://github.com/robots.txt
200
2274 bytes
5 User-agent groups
- /*?tab=achievements&achievement=*
- /*/*/pulse
- /*/*/projects
- /*/*/forks
- /*/*/issues/new
- /*/*/milestones/new
- /*/*/issues/search
- /*/*/commits/
- /*/*/branches
- /*/*/contributors
- /*/*/tags
- /*/*/stargazers
- /*/*/watchers
- /*/*/network
- /*/*/graphs
- /*/*/compare
- /*/tree/
- /gist/
- /*/download
- /*/revisions
- /*/commits/*?author
- /*/commits/*?path
- /*/comments
- /*/archive/
- /*/blame/
- /*/raw/
- /*/cache/
- /.git/
- */.git/
- /*.git$
- /search/advanced
- /search$
- /*q=
- /*.atom$
- /ekansa/Open-Context-Data
- /ekansa/opencontext-*
- */tarball/
- */zipball/
- /*source=*
- /*ref_cta=*
- /*plan=*
- /*return_to=*
- /*ref_loc=*
- /*setup_organization=*
- /*source_repo=*
- /*ref_page=*
- /*source=*
- /*referrer=*
- /*report=*
- /*author=*
- /*since=*
- /*until=*
- /*commits?author=*
- /*report-abuse?report=*
- /*tab=*
- /account-login
- /Explodingstuff/
- /copilot/
- /copilot/c/
- /*?tab=achievements&achievement=*
- /*/*/pulse
- /*/*/projects
- /*/*/forks
- /*/*/issues/new
- /*/*/milestones/new
- /*/*/issues/search
- /*/*/commits/
- /*/*/branches
- /*/*/contributors
- /*/*/tags
- /*/*/stargazers
- /*/*/watchers
- /*/*/network
- /*/*/graphs
- /*/*/compare
- /*/tree/
- /gist/
- /*/download
- /*/revisions
- /*/commits/*?author
- /*/commits/*?path
- /*/comments
- /*/archive/
- /*/blame/
- /*/raw/
- /*/cache/
- /.git/
- */.git/
- /*.git$
- /search/advanced
- /search$
- /*q=
- /*.atom$
- /ekansa/Open-Context-Data
- /ekansa/opencontext-*
- */tarball/
- */zipball/
- /*source=*
- /*ref_cta=*
- /*plan=*
- /*return_to=*
- /*ref_loc=*
- /*setup_organization=*
- /*source_repo=*
- /*ref_page=*
- /*source=*
- /*referrer=*
- /*report=*
- /*author=*
- /*since=*
- /*until=*
- /*commits?author=*
- /*report-abuse?report=*
- /*tab=*
- /account-login
- /Explodingstuff/
- /copilot/
- /copilot/c/
- /*?tab=achievements&achievement=*
- /*/*/pulse
- /*/*/projects
- /*/*/forks
- /*/*/issues/new
- /*/*/milestones/new
- /*/*/issues/search
- /*/*/commits/
- /*/*/branches
- /*/*/contributors
- /*/*/tags
- /*/*/stargazers
- /*/*/watchers
- /*/*/network
- /*/*/graphs
- /*/*/compare
- /*/tree/
- /gist/
- /*/download
- /*/revisions
- /*/commits/*?author
- /*/commits/*?path
- /*/comments
- /*/archive/
- /*/blame/
- /*/raw/
- /*/cache/
- /.git/
- */.git/
- /*.git$
- /search/advanced
- /search$
- /*q=
- /*.atom$
- /ekansa/Open-Context-Data
- /ekansa/opencontext-*
- */tarball/
- */zipball/
- /*source=*
- /*ref_cta=*
- /*plan=*
- /*return_to=*
- /*ref_loc=*
- /*setup_organization=*
- /*source_repo=*
- /*ref_page=*
- /*source=*
- /*referrer=*
- /*report=*
- /*author=*
- /*since=*
- /*until=*
- /*commits?author=*
- /*report-abuse?report=*
- /*tab=*
- /account-login
- /Explodingstuff/
- /copilot/
- /copilot/c/
- /*?tab=achievements&achievement=*
- /*/*/pulse
- /*/*/projects
- /*/*/forks
- /*/*/issues/new
- /*/*/milestones/new
- /*/*/issues/search
- /*/*/commits/
- /*/*/branches
- /*/*/contributors
- /*/*/tags
- /*/*/stargazers
- /*/*/watchers
- /*/*/network
- /*/*/graphs
- /*/*/compare
- /*/tree/
- /gist/
- /*/download
- /*/revisions
- /*/commits/*?author
- /*/commits/*?path
- /*/comments
- /*/archive/
- /*/blame/
- /*/raw/
- /*/cache/
- /.git/
- */.git/
- /*.git$
- /search/advanced
- /search$
- /*q=
- /*.atom$
- /ekansa/Open-Context-Data
- /ekansa/opencontext-*
- */tarball/
- */zipball/
- /*source=*
- /*ref_cta=*
- /*plan=*
- /*return_to=*
- /*ref_loc=*
- /*setup_organization=*
- /*source_repo=*
- /*ref_page=*
- /*source=*
- /*referrer=*
- /*report=*
- /*author=*
- /*since=*
- /*until=*
- /*commits?author=*
- /*report-abuse?report=*
- /*tab=*
- /account-login
- /Explodingstuff/
- /copilot/
- /copilot/c/
- /*?tab=achievements&achievement=*
- /*/*/pulse
- /*/*/projects
- /*/*/forks
- /*/*/issues/new
- /*/*/milestones/new
- /*/*/issues/search
- /*/*/commits/
- /*/*/branches
- /*/*/contributors
- /*/*/tags
- /*/*/stargazers
- /*/*/watchers
- /*/*/network
- /*/*/graphs
- /*/*/compare
- /*/tree/
- /gist/
- /*/download
- /*/revisions
- /*/commits/*?author
- /*/commits/*?path
- /*/comments
- /*/archive/
- /*/blame/
- /*/raw/
- /*/cache/
- /.git/
- */.git/
- /*.git$
- /search/advanced
- /search$
- /*q=
- /*.atom$
- /ekansa/Open-Context-Data
- /ekansa/opencontext-*
- */tarball/
- */zipball/
- /*source=*
- /*ref_cta=*
- /*plan=*
- /*return_to=*
- /*ref_loc=*
- /*setup_organization=*
- /*source_repo=*
- /*ref_page=*
- /*source=*
- /*referrer=*
- /*report=*
- /*author=*
- /*since=*
- /*until=*
- /*commits?author=*
- /*report-abuse?report=*
- /*tab=*
- /account-login
- /Explodingstuff/
- /copilot/
- /copilot/c/
Raw robots.txt
# If you would like to crawl GitHub contact us via https://support.github.com?tags=dotcom-robots # We also provide an extensive API: https://docs.github.com User-agent: bingbot Disallow: /ekansa/Open-Context-Data Disallow: /ekansa/opencontext-* Disallow: /account-login Disallow: */tarball/ Disallow: */zipball/ Disallow: /Explodingstuff/ Disallow: /copilot/ Disallow: /copilot/c/ User-agent: adidxbot Disallow: /ekansa/Open-Context-Data Disallow: /ekansa/opencontext-* Disallow: /account-login Disallow: */tarball/ Disallow: */zipball/ Disallow: /Explodingstuff/ Disallow: /copilot/ Disallow: /copilot/c/ User-agent: BingPreview Disallow: /ekansa/Open-Context-Data Disallow: /ekansa/opencontext-* Disallow: /account-login Disallow: */tarball/ Disallow: */zipball/ Disallow: /Explodingstuff/ Disallow: /copilot/ Disallow: /copilot/c/ User-agent: baidu crawl-delay: 1 User-agent: * Disallow: /*/*/pulse Disallow: /*/*/projects Disallow: /*/*/forks Disallow: /*/*/issues/new Disallow: /*/*/milestones/new Disallow: /*/*/issues/search Disallow: /*/*/commits/ Disallow: /*/*/branches Disallow: /*/*/contributors Disallow: /*/*/tags Disallow: /*/*/stargazers Disallow: /*/*/watchers Disallow: /*/*/network Disallow: /*/*/graphs Disallow: /*/*/compare Disallow: /*/tree/ Disallow: /gist/ Disallow: /*/download Disallow: /*/revisions Disallow: /*/commits/*?author Disallow: /*/commits/*?path Disallow: /*/comments Disallow: /*/archive/ Disallow: /*/blame/ Disallow: /*/raw/ Disallow: /*/cache/ Disallow: /.git/ Disallow: */.git/ Disallow: /*.git$ Disallow: /search/advanced Disallow: /search$ Disallow: /*q= Disallow: /*.atom$ Disallow: /ekansa/Open-Context-Data Disallow: /ekansa/opencontext-* Disallow: */tarball/ Disallow: */zipball/ Disallow: /*source=* Disallow: /*ref_cta=* Disallow: /*plan=* Disallow: /*return_to=* Disallow: /*ref_loc=* Disallow: /*setup_organization=* Disallow: /*source_repo=* Disallow: /*ref_page=* Disallow: /*source=* Disallow: /*referrer=* Disallow: /*report=* Disallow: /*author=* Disallow: /*since=* Disallow: /*until=* Disallow: /*commits?author=* Disallow: /*report-abuse?report=* Disallow: /*tab=* Allow: /*?tab=achievements&achievement=* Disallow: /account-login Disallow: /Explodingstuff/ Disallow: /copilot/ Disallow: /copilot/c/
-
1
Paste your input
Enter the value at the top — domain, IP, URL, email, ASN, hash, whatever fits this tool. The smart input auto-detects type.
-
2
Click "Inspect"
host.tools issues real probes (DNS, HTTP, TCP, TLS, WHOIS where applicable) and renders the result in milliseconds.
-
3
Open the API tab
Every web tool has a sibling /api/v1/http/robots JSON endpoint with the same payload. One copy-as-curl click and you're scripting it.
Headers are how the modern web declares its security posture. Auditing them is the highest-ROI thing you can do this week.
/api/v1/http/robots?q=https%3A%2F%2Fgithub.com
curl -s '/api/v1/http/robots?q=https%3A%2F%2Fgithub.com'
<iframe src="/http/robots?q={INPUT}&embed=1"
width="100%" height="600" frameborder="0"></iframe>
Drop into any HTML page. The embed=1 flag hides nav and footer.
Upgrade to Pro for $19/mo. Cancel anytime. Works with the same API you already use.
Common questions
Is robots.txt parser free?
Where does the data come from?
How fresh are the results?
?nocache=1 for a forced refresh.Can I run this from the command line?
host.tools http robots YOUR_INPUT.Can I monitor results over time?
Run robots.txt parser on a schedule. Get pinged when it changes.
Pro gets you bulk lookups, monitors, webhook alerts, history, exports and 10,000 API calls/day. $19/mo.
- ✓Schedule any tool — every 1, 5, 15, 60 min
- ✓Diff against last run, alert on change
- ✓Webhook + email + Slack + PagerDuty + OpsGenie
- ✓Bulk CSV upload, 1,000 inputs per job
- ✓Export results as CSV / NDJSON / Excel
- ✓90-day history, comparison view