Metadata-Version: 2.1
Name: datasets
Version: 1.6.2
Summary:  HuggingFace/Datasets is an open library of NLP datasets.
Home-page: https://github.com/huggingface/datasets
Author: HuggingFace Inc.
Author-email: thomas@huggingface.co
License: Apache 2.0
Download-URL: https://github.com/huggingface/datasets/tags
Description: Note:
        
           VERSION needs to be formatted following the MAJOR.MINOR.PATCH convention
           (we need to follow this convention to be able to retrieve versioned scripts)
        
        Simple check list for release from AllenNLP repo: https://github.com/allenai/allennlp/blob/master/setup.py
        
        To create the package for pypi.
        
        1. Change the version in __init__.py, setup.py as well as docs/source/conf.py.
        
        2. Commit these changes with the message: "Release: VERSION"
        
        3. Add a tag in git to mark the release: "git tag VERSION -m'Adds tag VERSION for pypi' "
           Push the tag to git: git push --tags origin master
        
        4. Build both the sources and the wheel. Do not change anything in setup.py between
           creating the wheel and the source distribution (obviously).
        
           For the wheel, run: "python setup.py bdist_wheel" in the top level directory.
           (this will build a wheel for the python version you use to build it).
        
           For the sources, run: "python setup.py sdist"
           You should now have a /dist directory with both .whl and .tar.gz source versions.
        
        5. Check that everything looks correct by uploading the package to the pypi test server:
        
           twine upload dist/* -r pypitest
           (pypi suggest using twine as other methods upload files via plaintext.)
           You may have to specify the repository url, use the following command then:
           twine upload dist/* -r pypitest --repository-url=https://test.pypi.org/legacy/
        
           Check that you can install it in a virtualenv by running:
           pip install -i https://testpypi.python.org/pypi datasets
        
        6. Upload the final version to actual pypi:
           twine upload dist/* -r pypi
        
        7. Fill release notes in the tag in github once everything is looking hunky-dory.
        
        8. Update the documentation commit in .circleci/deploy.sh for the accurate documentation to be displayed
           Update the version mapping in docs/source/_static/js/custom.js,
           and set version to X.X.X.dev0 in setup.py and __init__.py
        
        
Keywords: datasets machine learning datasets metrics
Platform: UNKNOWN
Classifier: Development Status :: 5 - Production/Stable
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: Education
Classifier: Intended Audience :: Science/Research
Classifier: License :: OSI Approved :: Apache Software License
Classifier: Operating System :: OS Independent
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.6
Classifier: Programming Language :: Python :: 3.7
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Provides-Extra: apache-beam
Provides-Extra: tensorflow
Provides-Extra: tensorflow_gpu
Provides-Extra: torch
Provides-Extra: s3
Provides-Extra: dev
Provides-Extra: tests
Provides-Extra: quality
Provides-Extra: benchmarks
Provides-Extra: docs
