Skip to content
This repository has been archived by the owner on Mar 9, 2021. It is now read-only.

Prevent spider from crawling certain directories or files #67

Open
seezee opened this issue May 30, 2019 · 0 comments
Open

Prevent spider from crawling certain directories or files #67

seezee opened this issue May 30, 2019 · 0 comments

Comments

@seezee
Copy link

seezee commented May 30, 2019

I've got some PHP files to prevent scrapers & form spammers. If a spider ignores my robots.txt file they are blacklisted. Currently, glyphhanger doesn't respect robots.txt and tries to crawl the directories or files.

An option like --spider-nocrawl=/wp-content/uploads/forbiddenfolder&&/wp-content/uploads/forbidden-file.php would be nice!

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant