Skip to main content

Scrapy middleware to add extra "magic" fields to items

Project description

https://travis-ci.org/scrapy-plugins/scrapy-magicfields.svg?branch=master https://codecov.io/gh/scrapy-plugins/scrapy-magicfields/branch/master/graph/badge.svg

This is a Scrapy spider middleware to add extra fields to items, based on the configuration settings MAGIC_FIELDS and MAGIC_FIELDS_OVERRIDE.

Installation

Install scrapy-magicfields using pip:

$ pip install scrapy-magicfields

Configuration

  1. Add MagicFieldsMiddleware by including it in SPIDER_MIDDLEWARES in your settings.py file:

    SPIDER_MIDDLEWARES = {
        'scrapy_magicfields.MagicFieldsMiddleware': 100,
    }

    Here, priority 100 is just an example. Set its value depending on other middlewares you may have enabled already.

  2. Enable the middleware using MAGIC_FIELDS (and optionally MAGIC_FIELDS_OVERRIDE) in your setting.py.

Usage

Both settings MAGIC_FIELDS and MAGIC_FIELDS_OVERRIDE are dicts:

  • the keys are the destination field names,

  • their value is a string which accepts magic variables, — identified by a starting $ (dollar sign), which will be substituted by a corresponding value at runtime.

Some magic variables also accept arguments, and are specified after the magic name, using a : (column) as separator.

You can set project-global magics with MAGIC_FIELDS, and tune them for a specific spider using MAGIC_FIELDS_OVERRIDE.

In case there is more than one argument, they must come separated by , (comma sign). So the generic magic format is:

$<magic name>[:arg1,arg2,...]

Supported magic variables

$time

the UTC timestamp at which the item was scraped, in format '%Y-%m-%d %H:%M:%S'.

$unixtime

the unixtime (number of seconds since the Epoch, i.e. time.time()) at which the item was scraped.

$isotime

the UTC timestamp at which the item was scraped, with format '%Y-%m-%dT%H:%M:%S".

$spider

must be followed by an argument, which is the name of an attribute of the spider (like an argument passed to it).

$env

the value of an environment variable. It acccepts as argument the name of the variable.

$jobid

the job id (shortcut for $env:SCRAPY_JOB)

$jobtime

the UTC timestamp at which the job started, in format '%Y-%m-%d %H:%M:%S'.

$response

Access to some response properties.

$response:url

The url from where the item was extracted from.

$response:status

Response http status.

$response:headers

Response http headers.

$setting

Access the given Scrapy setting. It accepts one argument: the name of the setting.

$field

Allows to copy the value of one field to another Its argument is the source field. Effects are unpredicable if you use as source a field that is filled using magic fields.

Examples

The following configuration will add two fields to each scraped item:

  • 'timestamp', which will be filled with the string 'item scraped at <scraped timestamp>',

  • and 'spider', which will contain the spider name

MAGIC_FIELDS = {
    "timestamp": "item scraped at $time",
    "spider": "$spider:name"
}

The following configuration will copy the url to the field sku:

MAGIC_FIELDS = {
    "sku": "$field:url"
}

Magics also accept a regular expression argument which allows to extract and assign only part of the value generated by the magic. You have to specify it using the r'' notation.

Let’s pretend that the urls of your items look like 'http://www.example.com/product.html?item_no=345' and you want to assign to the sku field only the item number.

The following example, similar to the previous one but with a second regular expression argument, will do the task:

MAGIC_FIELDS = {
    "sku": "$field:url,r'item_no=(\d+)'"
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

scrapy-magicfields-1.0.0.tar.gz (3.9 kB view details)

Uploaded Source

Built Distribution

scrapy_magicfields-1.0.0-py2.py3-none-any.whl (3.9 kB view details)

Uploaded Python 2 Python 3

File details

Details for the file scrapy-magicfields-1.0.0.tar.gz.

File metadata

File hashes

Hashes for scrapy-magicfields-1.0.0.tar.gz
Algorithm Hash digest
SHA256 18fe0cba1411dacbad2d6b85b337fd7edd540adeb247dbf7dbad40635e554913
MD5 f431f1b8728748f40caa5b13f3d2cc8c
BLAKE2b-256 c8a83c54cabd3e0bace74b4be3f4db3dad8ab67f1ca44d4fd3d8da7d7e29de21

See more details on using hashes here.

File details

Details for the file scrapy_magicfields-1.0.0-py2.py3-none-any.whl.

File metadata

File hashes

Hashes for scrapy_magicfields-1.0.0-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 52f26f41628cf5be80d500865d041ad2fee1fed2a8d530085a8b6845fe5aeb13
MD5 71973a8e9062b558a0e3b1423bc9d4ff
BLAKE2b-256 b51430b44e6f5a66fe7bd5ccb1e367e1aa4e9acba63af77539436504a2ccd8c5

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page