File size: 5,228 Bytes
626eca0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 |
"""
Simple check list from AllenNLP repo: https://github.com/allenai/allennlp/blob/master/setup.py
To create the package for pypi.
1. Run `make pre-release` (or `make pre-patch` for a patch release) then run `make fix-copies` to fix the index of the
documentation.
2. Run Tests for Amazon Sagemaker. The documentation is located in `./tests/sagemaker/README.md`, otherwise @philschmid.
3. Unpin specific versions from setup.py that use a git install.
4. Commit these changes with the message: "Release: VERSION"
5. Add a tag in git to mark the release: "git tag VERSION -m 'Adds tag VERSION for pypi' "
Push the tag to git: git push --tags origin master
6. Build both the sources and the wheel. Do not change anything in setup.py between
creating the wheel and the source distribution (obviously).
For the wheel, run: "python setup.py bdist_wheel" in the top level directory.
(this will build a wheel for the python version you use to build it).
For the sources, run: "python setup.py sdist"
You should now have a /dist directory with both .whl and .tar.gz source versions.
7. Check that everything looks correct by uploading the package to the pypi test server:
twine upload dist/* -r pypitest
(pypi suggest using twine as other methods upload files via plaintext.)
You may have to specify the repository url, use the following command then:
twine upload dist/* -r pypitest --repository-url=https://test.pypi.org/legacy/
Check that you can install it in a virtualenv by running:
pip install -i https://testpypi.python.org/pypi transformers
8. Upload the final version to actual pypi:
twine upload dist/* -r pypi
9. Copy the release notes from RELEASE.md to the tag in github once everything is looking hunky-dory.
10. Run `make post-release` (or `make post-patch` for a patch release).
"""
from collections import defaultdict
import setuptools
def parse_requirements_file(
path, allowed_extras: set = None, include_all_extra: bool = True
):
requirements = []
extras = defaultdict(list)
find_links = []
with open(path) as requirements_file:
import re
def fix_url_dependencies(req: str) -> str:
"""Pip and setuptools disagree about how URL dependencies should be handled."""
m = re.match(
r"^(git\+)?(https|ssh)://(git@)?github\.com/([\w-]+)/(?P<name>[\w-]+)\.git",
req,
)
if m is None:
return req
else:
return f"{m.group('name')} @ {req}"
for line in requirements_file:
line = line.strip()
if line.startswith("#") or len(line) <= 0:
continue
if (
line.startswith("-f")
or line.startswith("--find-links")
or line.startswith("--index-url")
):
find_links.append(line.split(" ", maxsplit=1)[-1].strip())
continue
req, *needed_by = line.split("# needed by:")
req = fix_url_dependencies(req.strip())
if needed_by:
for extra in needed_by[0].strip().split(","):
extra = extra.strip()
if allowed_extras is not None and extra not in allowed_extras:
raise ValueError(f"invalid extra '{extra}' in {path}")
extras[extra].append(req)
if include_all_extra and req not in extras["all"]:
extras["all"].append(req)
else:
requirements.append(req)
return requirements, extras, find_links
allowed_extras = {
"onnx",
"onnx-gpu",
"serve",
"retriever",
"reader",
"all",
"faiss",
"dev",
}
# Load requirements.
install_requirements, extras, find_links = parse_requirements_file(
"requirements.txt", allowed_extras=allowed_extras
)
# version.py defines the VERSION and VERSION_SHORT variables.
# We use exec here, so we don't import allennlp whilst setting up.
VERSION = {} # type: ignore
with open("relik/version.py", "r") as version_file:
exec(version_file.read(), VERSION)
with open("README.md", "r") as fh:
long_description = fh.read()
setuptools.setup(
name="relik",
version=VERSION["VERSION"],
author="Edoardo Barba, Riccardo Orlando, Pere-Lluís Huguet Cabot",
author_email="[email protected]",
description="Fast and Accurate Entity Linking and Relation Extraction on an Academic Budget",
long_description=long_description,
long_description_content_type="text/markdown",
url="https://github.com/SapienzaNLP/relik",
keywords="NLP Sapienza sapienzanlp deep learning transformer pytorch retriever entity linking relation extraction reader budget",
packages=setuptools.find_packages(),
include_package_data=True,
license="Apache",
classifiers=[
"Intended Audience :: Science/Research",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python :: 3",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
],
install_requires=install_requirements,
extras_require=extras,
python_requires=">=3.10",
)
|