Skip to content

Basic Usage

First steps

If you already have the Crawler installed, basic startup is trivial.

From the folder where you installed/extracted the Crawler, you can run the following commands.

Crawl a website and print the results to the console

Terminal window
./crawler --url=

Even this basic command will analyze the specified page and save the report in various formats to tmp/ folder (by default):

  • Directorytmp/

Crawl with HTML report to e-mail

./crawler \
--url= \ \

Generate offline version of the website

./crawler \
--url= \

After executing this command, you can open the tmp/ file in your browser and you will be able to view and browse the whole website without internet connection.

Lot of other uses