Skip to main content

DataLad extension package for crawling external web resources into an automated data distribution

Project description

 ____          _           _                 _
|  _ \   __ _ | |_   __ _ | |      __ _   __| |
| | | | / _` || __| / _` || |     / _` | / _` |
| |_| || (_| || |_ | (_| || |___ | (_| || (_| |
|____/  \__,_| \__| \__,_||_____| \__,_| \__,_|
                                   Crawler

Travis tests status codecov.io Documentation License: MIT GitHub release PyPI version fury.io Average time to resolve an issue Percentage of issues still open

This extension enhances DataLad (http://datalad.org) for crawling external web resources into an automated data distribution. Please see the extension documentation for a description on additional commands and functionality.

For general information on how to use or contribute to DataLad (and this extension), please see the DataLad website or the main GitHub project page.

Installation

Before you install this package, please make sure that you install a recent version of git-annex. Afterwards, install the latest version of datalad-crawler from PyPi. It is recommended to use a dedicated virtualenv:

# create and enter a new virtual environment (optional)
virtualenv --system-site-packages --python=python3 ~/env/datalad
. ~/env/datalad/bin/activate

# install from PyPi
pip install datalad_crawler

Support

The documentation of this project is found here: http://docs.datalad.org/projects/crawler

All bugs, concerns and enhancement requests for this software can be submitted here: https://github.com/datalad/datalad-crawler/issues

If you have a problem or would like to ask a question about how to use DataLad, please submit a question to NeuroStars.org with a datalad tag. NeuroStars.org is a platform similar to StackOverflow but dedicated to neuroinformatics.

All previous DataLad questions are available here: http://neurostars.org/tags/datalad/

Acknowledgements

DataLad development is supported by a US-German collaboration in computational neuroscience (CRCNS) project "DataGit: converging catalogues, warehouses, and deployment logistics into a federated 'data distribution'" (Halchenko/Hanke), co-funded by the US National Science Foundation (NSF 1429999) and the German Federal Ministry of Education and Research (BMBF 01GQ1411). Additional support is provided by the German federal state of Saxony-Anhalt and the European Regional Development Fund (ERDF), Project: Center for Behavioral Brain Sciences, Imaging Platform. This work is further facilitated by the ReproNim project (NIH 1P41EB019936-01A1).

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

datalad_crawler-0.9.4.tar.gz (111.5 kB view details)

Uploaded Source

Built Distribution

datalad_crawler-0.9.4-py3-none-any.whl (146.8 kB view details)

Uploaded Python 3

File details

Details for the file datalad_crawler-0.9.4.tar.gz.

File metadata

  • Download URL: datalad_crawler-0.9.4.tar.gz
  • Upload date:
  • Size: 111.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.8.0 pkginfo/1.8.2 readme-renderer/34.0 requests/2.27.1 requests-toolbelt/0.9.1 urllib3/1.26.9 tqdm/4.64.0 importlib-metadata/4.8.3 keyring/23.4.1 rfc3986/1.5.0 colorama/0.4.4 CPython/3.6.15

File hashes

Hashes for datalad_crawler-0.9.4.tar.gz
Algorithm Hash digest
SHA256 53e1a8ec201d4bd7c0c567dd312af769270d73f22cd8613f48570ed45aea437e
MD5 173aea00a03fd99e8c28dbc3a6920a38
BLAKE2b-256 c718d5b32312d9f1bc8705316af95bbb46bb804f6ab67f31ddb70e5ce2e73b66

See more details on using hashes here.

Provenance

File details

Details for the file datalad_crawler-0.9.4-py3-none-any.whl.

File metadata

  • Download URL: datalad_crawler-0.9.4-py3-none-any.whl
  • Upload date:
  • Size: 146.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.8.0 pkginfo/1.8.2 readme-renderer/34.0 requests/2.27.1 requests-toolbelt/0.9.1 urllib3/1.26.9 tqdm/4.64.0 importlib-metadata/4.8.3 keyring/23.4.1 rfc3986/1.5.0 colorama/0.4.4 CPython/3.6.15

File hashes

Hashes for datalad_crawler-0.9.4-py3-none-any.whl
Algorithm Hash digest
SHA256 e5e686c7b93653d0711cd981ad2d7e8ccece148d197292b580c3012db3d5d32d
MD5 3aa2725254b2ff9651e02db5298f89d3
BLAKE2b-256 f4a50f55adbaf012ff815b4e22c3047706404da75daa9a145d5f3fd6978c51e4

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page