Skip to content

Crawl pages within a domain, reporting any page that returns a bad response code

License

Notifications You must be signed in to change notification settings

alphasights/crawl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

68 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Crawl

Circle CI

Crawl pages within a domain, reporting any page that returns a bad response code

Usage:

> crawl [options] domain

Usage: crawl [options] domain
-s, --start /home,/about         Starting path(s), defaults to /
-u, --username username          Basic auth username
-p, --password password          Basic auth password
-c, --connections count          Max mumber of parallel connections to use. The default is 5.
-v, --verbose                    Give details when crawling
-h, --help                       Show this message
    --version                    Print version

Example:

> crawl https://engineering.alphasights.com --connections=5 --start=/ --verbose

  Adding /
Fetching / ...
  Adding /positions/ruby-developer
  Adding /positions/js-ember-developer
  Adding /positions/ux-ui-designer
  Adding /positions/support-specialist
Fetching /positions/ruby-developer
Fetching /positions/js-ember-developer ...
Fetching /positions/ux-ui-designer ...
Fetching /positions/support-specialist ...

5 pages crawled without errors.

Copyright and License

Copyright AlphaSights and Contributors, 2015

MIT Licence

About

Crawl pages within a domain, reporting any page that returns a bad response code

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages