Skip to main content

HuggingFace/Datasets is an open library of NLP datasets.

Project description

Note:

VERSION needs to be formatted following the MAJOR.MINOR.PATCH convention (we need to follow this convention to be able to retrieve versioned scripts)

Simple check list for release from AllenNLP repo: https://github.com/allenai/allennlp/blob/master/setup.py

To create the package for pypi.

  1. Change the version in __init__.py, setup.py as well as docs/source/conf.py.

  2. Commit these changes with the message: “Release: VERSION”

  3. Add a tag in git to mark the release: “git tag VERSION -m’Adds tag VERSION for pypi’ ” Push the tag to git: git push –tags origin master

  4. Build both the sources and the wheel. Do not change anything in setup.py between creating the wheel and the source distribution (obviously).

    First pin the SCRIPTS_VERSION to VERSION in __init__.py (but don’t commit this change)

    For the wheel, run: “python setup.py bdist_wheel” in the top level directory. (this will build a wheel for the python version you use to build it).

    For the sources, run: “python setup.py sdist” You should now have a /dist directory with both .whl and .tar.gz source versions.

    Then change the SCRIPTS_VERSION back to to “master” in __init__.py (but don’t commit this change)

  5. Check that everything looks correct by uploading the package to the pypi test server:

    twine upload dist/* -r pypitest (pypi suggest using twine as other methods upload files via plaintext.) You may have to specify the repository url, use the following command then: twine upload dist/* -r pypitest –repository-url=https://test.pypi.org/legacy/

    Check that you can install it in a virtualenv by running: pip install -i https://testpypi.python.org/pypi datasets

  6. Upload the final version to actual pypi: twine upload dist/* -r pypi

  7. Copy the release notes from RELEASE.md to the tag in github once everything is looking hunky-dory.

  8. Update the documentation commit in .circleci/deploy.sh for the accurate documentation to be displayed Update the version mapping in docs/source/_static/js/custom.js

  9. Update README.md to redirect to correct documentation.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

datasets-1.1.3.tar.gz (131.5 kB view details)

Uploaded Source

Built Distribution

datasets-1.1.3-py3-none-any.whl (153.7 kB view details)

Uploaded Python 3

File details

Details for the file datasets-1.1.3.tar.gz.

File metadata

  • Download URL: datasets-1.1.3.tar.gz
  • Upload date:
  • Size: 131.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/46.1.3 requests-toolbelt/0.9.1 tqdm/4.45.0 CPython/3.7.2

File hashes

Hashes for datasets-1.1.3.tar.gz
Algorithm Hash digest
SHA256 40261f45806ebe003194bb6d14b3f59a6f1e7f9e347e78b662e1ab979ace7e9c
MD5 8b8bea8b0de57815c1e4b4dd9e819ca0
BLAKE2b-256 3808e6e94005e882d61780810b180f1969871f5df0583e0700c3ae189581eaaa

See more details on using hashes here.

File details

Details for the file datasets-1.1.3-py3-none-any.whl.

File metadata

  • Download URL: datasets-1.1.3-py3-none-any.whl
  • Upload date:
  • Size: 153.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/46.1.3 requests-toolbelt/0.9.1 tqdm/4.45.0 CPython/3.7.2

File hashes

Hashes for datasets-1.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 2c5bbd3abd563da4d770c26de4296ed8b6033837f46dc833f4bcdf893aa4ffbc
MD5 a64e6853d5b20813d64e17dcd7c5c456
BLAKE2b-256 1a380c24dce24767386123d528d27109024220db0e7a04467b658d587695241a

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page