-
Notifications
You must be signed in to change notification settings - Fork 38
/
Copy pathsetup.py
60 lines (52 loc) · 2.04 KB
/
setup.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
import subprocess
import sys
from setuptools.command.install import install
from setuptools import setup, find_packages
with open('requirements.txt') as f:
requirements = f.read().splitlines()
tests_require = ['pytest', 'pytest-cov', 'coverage']
with open("docs/ETL_README.md", "r") as f:
long_description = f.read()
class ShellInstall(install):
def run(self):
if not sys.platform.startswith("linux"):
print('Your platform {} might not be supported'.format(sys.platform))
else:
print('Running create_python_venv.sh -n hello-fresh-data-engg')
subprocess.call(['./sbin/create_python_venv.sh', '-n', 'hello-fresh-data-engg'])
install.run(self)
setup(
cmdclass={'install': ShellInstall},
name='datapipelines-essentials',
version='2.0',
author='Vitthal Mirji',
author_email='[email protected]',
url='https://vitthalmirji.com',
description='Datalake complex transformations simplified in PySpark',
long_description='Simplified ETL process in Hadoop using Apache Spark. '
'SparkSession extensions, DataFrame validation, Column extensions, SQL functions, and DataFrame '
'transformations',
long_description_content_type="text/markdown",
install_requires=requirements,
tests_require=tests_require,
extras_require={
'test': tests_require,
'all': requirements + tests_require,
'docs': ['sphinx'] + tests_require,
'lint': []
},
license="GNU :: GPLv3",
include_package_data=True,
packages=find_packages(where='src', include=['com*']),
package_dir={"": "src"},
setup_requires=['setuptools'],
classifiers=[
"Programming Language :: Python :: 3",
"License :: GNU :: GPLv3",
"Operating System :: Linux",
],
dependency_links=[],
python_requires='>=3.7,<=3.9.5',
keywords=['apachespark', 'spark', 'pyspark', 'etl', 'hadoop', 'bigdata', 'apache-spark', 'python', 'python3',
'data', 'dataengineering', 'datapipelines']
)