|
16 | 16 |
|
17 | 17 | install_requires = io.open("./requirements.txt", encoding="utf-8").read() |
18 | 18 |
|
| 19 | +core_requires = io.open("./requirements.txt", encoding="utf-8").read() |
19 | 20 |
|
20 | 21 | test_requires = io.open("./test/requirements.txt", encoding="utf-8").read() |
21 | 22 |
|
22 | | -azure_identity = 'azure-identity<1.5.0' |
23 | | -azure_storage_file_datalake = "azure-storage-file-datalake<=12.4.0" |
24 | | -azureml_core = "azureml-core>=1.21.0.post2,<=1.21.0.post2" |
25 | | -commondatamodel_objectmodel = "commondatamodel-objectmodel>=1.2.2,<=1.2.2" |
26 | | -dotenv = "python-dotenv>=0.14.0,<=0.14.0" |
27 | | -ms_recommenders = "ms-recommenders>=0.6.0,<=0.6.0" |
28 | | -pandas = "pandas>=1.0.3,<=1.1.3" |
29 | | -pydantic = "pydantic>1.7.3,<=1.7.4" |
30 | | -pyspark = "pyspark>2.4.2,<=2.4.5" |
31 | | -requests = "requests>=2.24.0,<=2.24.0" |
32 | | - |
33 | | -CI_RETAIL_UTILS = [ |
34 | | - azure_storage_file_datalake, |
35 | | - azureml_core, |
36 | | - commondatamodel_objectmodel, |
37 | | - ms_recommenders, |
38 | | - pydantic, |
39 | | - pyspark, |
40 | | - dotenv, |
41 | | - requests |
42 | | -] |
43 | | - |
44 | | -CI_CDM2AI = [ |
45 | | - azure_identity, |
46 | | - azure_storage_file_datalake, |
47 | | - pandas, |
48 | | - pyspark, |
49 | | -] |
50 | | - |
51 | 23 | extras = { |
52 | 24 | "required": install_requires, |
53 | | - "all": install_requires, |
| 25 | + "core": core_requires, |
54 | 26 | "test": test_requires, |
55 | | - "retail-utils": CI_RETAIL_UTILS, |
56 | | - "retail-cdm2ai": CI_CDM2AI |
57 | 27 | } |
58 | 28 |
|
59 | 29 | SETUP_REQUIRES = [ |
|
67 | 37 |
|
68 | 38 | setup( |
69 | 39 | name="ai-python", |
70 | | - version="0.0.4", |
| 40 | + version="0.1.0", |
71 | 41 | description="Microsoft AI Python Package", |
72 | 42 | long_description=readme, |
73 | 43 | long_description_content_type="text/x-rst", |
|
0 commit comments