Eek, a [web] spider.
Project description
Eek is a web crawler that outputs metadata about a website in CSV format.
Installation
$ pip install eek
Usage
usage: eek [-h] [–graph] [–delay SECONDS] [–grep PATTERN] [-i] URL
eek recursively crawls a website, outputing metadata about each page in CSV format.
positional arguments: URL The base URL to start the crawl optional arguments: -h, --help show this help message and exit --graph output a graphviz digraph of links instead of CSV metadata --delay SECONDS Time, in seconds, to wait in between fetches. Defaults to 0. --grep PATTERN Print urls containing PATTERN (a python regular expression). -i, --ignore-case Ignore case. Only valid with --grep
Example:
eek http://example.com/
To save output to a file, use redirection
eek http://example.com/ > ~/some_file.csv
To slow down crawling, use --delay=[seconds]
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
eek-1.0.2.tar.gz
(7.4 kB
view details)
File details
Details for the file eek-1.0.2.tar.gz
.
File metadata
- Download URL: eek-1.0.2.tar.gz
- Upload date:
- Size: 7.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6e8532e0fbbdbf555c4ac7589d10bad78351d9ea9f8e421004997c3cb7dce198 |
|
MD5 | 4d0a795c358e3269b5add8613bb8e020 |
|
BLAKE2b-256 | 77a5fb62b6a3d85972018ebef8525016634a8559a7af5094a3c2186c27926ee1 |