Scroll to navigation

SCRAPY(1) General Commands Manual SCRAPY(1)

NAME

scrapy - the Scrapy command-line tool

SYNOPSIS

scrapy [command] [OPTIONS] ...

DESCRIPTION

Scrapy is controlled through the scrapy command-line tool. The script provides several commands, for different purposes. Each command supports its own particular syntax. In other words, each command supports a different set of arguments and options.

OPTIONS

fetch [OPTION] URL

Print response HTTP headers instead of body

runspider [OPTION] spiderfile

Run a spider

Store scraped items to FILE in XML format

settings [OPTION]

Query Scrapy settings

Print raw setting value
Print setting value, interpreted as a boolean
Print setting value, interpreted as an integer
Print setting value, interpreted as an float
Print setting value, interpreted as an float
Print initial setting value (before loading extensions and spiders)

shell URL | file

Launch the interactive scraping console

startproject projectname

Create new project with an initial project template

--help, -h

Print command help and options

--logfile=FILE

Log file. if omitted stderr will be used

--loglevel=LEVEL, -L LEVEL

Log level (default: None)

--nolog

Disable logging completely

--spider=SPIDER

Always use this spider when arguments are urls

--profile=FILE

Write python cProfile stats to FILE

--lsprof=FILE

Write lsprof profiling stats to FILE

--pidfile=FILE

Write process ID to FILE

--set=NAME=VALUE, -s NAME=VALUE

Set/override setting (may be repeated)

AUTHOR

Scrapy was written by the Scrapy Developers.

This manual page was written by Ignace Mouzannar <mouzannar@gmail.com>, for the Debian project (but may be used by others).

October 17, 2009