Skip to content

Commit 71e6c3b

Browse files
authored
format filues (#212)
1 parent 94c5132 commit 71e6c3b

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

73 files changed

+1027
-595
lines changed
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
"""Bank Subscription Prediction Project using ZenML."""
1+
"""Bank Subscription Prediction Project using ZenML."""
Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +0,0 @@
1-

bank_subscription_prediction/pipelines/training_pipeline.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,12 @@
1-
from zenml import pipeline
2-
from steps.data_loader import load_data
1+
import logging
2+
33
from steps.data_cleaner import clean_data_step
4+
from steps.data_loader import load_data
45
from steps.data_preprocessor import preprocess_data_step
56
from steps.data_splitter import split_data_step
6-
from steps.model_trainer import train_xgb_model_with_feature_selection
77
from steps.model_evaluator import evaluate_model
8-
import logging
8+
from steps.model_trainer import train_xgb_model_with_feature_selection
9+
from zenml import pipeline
910

1011
# Set up logger
1112
logger = logging.getLogger(__name__)

bank_subscription_prediction/predict_bank_cd_subs_by_xgboost_clf_for_imbalance_dataset.ipynb

Lines changed: 19 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -211,10 +211,9 @@
211211
"## 1. reading various files into the dataframe\n",
212212
"## 2. to performa various data manipulation tasks\n",
213213
"\n",
214-
"import pandas as pd\n",
215-
"\n",
216214
"# load numpy\n",
217-
"import numpy as np"
215+
"import numpy as np\n",
216+
"import pandas as pd"
218217
]
219218
},
220219
{
@@ -238,33 +237,27 @@
238237
"# for custom transformer\n",
239238
"from sklearn.base import BaseEstimator, TransformerMixin\n",
240239
"\n",
241-
"# for creating pipeline\n",
242-
"from sklearn.pipeline import Pipeline, FeatureUnion\n",
243-
"\n",
244240
"# for cross validation\n",
245241
"from sklearn.cross_validation import train_test_split\n",
246-
"from sklearn.cross_validation import train_test_split\n",
247-
"from sklearn import cross_validation\n",
248-
"from sklearn.model_selection import cross_val_predict, cross_val_score\n",
249-
"\n",
250-
"# for various metrics and reporting\n",
251-
"from sklearn import metrics\n",
252-
"from sklearn.metrics import roc_curve\n",
253-
"from sklearn.metrics import precision_recall_curve\n",
254-
"from sklearn.metrics import accuracy_score\n",
255-
"from sklearn.metrics import confusion_matrix\n",
256-
"from sklearn.metrics import classification_report\n",
257-
"from sklearn.metrics import accuracy_score\n",
258-
"from sklearn.metrics import confusion_matrix\n",
259242
"\n",
260243
"# feature selection\n",
261244
"from sklearn.feature_selection import SelectFromModel\n",
262245
"\n",
263-
"# xgboost library\n",
264-
"from xgboost import XGBClassifier\n",
246+
"# for various metrics and reporting\n",
247+
"from sklearn.metrics import (\n",
248+
" classification_report,\n",
249+
" confusion_matrix,\n",
250+
" precision_recall_curve,\n",
251+
" roc_curve,\n",
252+
")\n",
253+
"from sklearn.model_selection import cross_val_predict\n",
265254
"\n",
255+
"# for creating pipeline\n",
256+
"from sklearn.pipeline import FeatureUnion, Pipeline\n",
257+
"\n",
258+
"# xgboost library\n",
266259
"# plot feature importance\n",
267-
"from xgboost import plot_importance, plot_tree"
260+
"from xgboost import XGBClassifier, plot_importance"
268261
]
269262
},
270263
{
@@ -313,11 +306,11 @@
313306
"# following code is for IBM Watson Studio loading. note if you are not using Watson studion ,\n",
314307
"# one can skip this.\n",
315308
"\n",
316-
"import sys\n",
317309
"import types\n",
310+
"\n",
311+
"import ibm_boto3\n",
318312
"import pandas as pd\n",
319313
"from botocore.client import Config\n",
320-
"import ibm_boto3\n",
321314
"\n",
322315
"\n",
323316
"def __iter__(self):\n",
@@ -1248,10 +1241,10 @@
12481241
"# Definition of the CategoricalEncoder class, copied from PR #9151.\n",
12491242
"# Just run this cell, or copy it to your code, do not try to understand it (yet).\n",
12501243
"\n",
1244+
"from scipy import sparse\n",
12511245
"from sklearn.base import BaseEstimator, TransformerMixin\n",
1252-
"from sklearn.utils import check_array\n",
12531246
"from sklearn.preprocessing import LabelEncoder\n",
1254-
"from scipy import sparse\n",
1247+
"from sklearn.utils import check_array\n",
12551248
"\n",
12561249
"\n",
12571250
"class CategoricalEncoder(BaseEstimator, TransformerMixin):\n",

bank_subscription_prediction/run.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,8 @@
11
"""Main module to run the Bank Subscription Prediction pipeline."""
22

3-
import click
43
import logging
4+
5+
import click
56
from pipelines.training_pipeline import bank_subscription_training_pipeline
67

78
logger = logging.getLogger(__name__)
Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +0,0 @@
1-

bank_subscription_prediction/steps/data_loader.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,13 @@
1-
import pandas as pd
1+
import logging
22
import os
3-
import requests
4-
from zenml import step
53

64
# Download the zip file
75
import zipfile
86
from io import BytesIO
9-
import logging
7+
8+
import pandas as pd
9+
import requests
10+
from zenml import step
1011

1112
# Set up logger
1213
logger = logging.getLogger(__name__)

bank_subscription_prediction/steps/data_preprocessor.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,8 @@
1+
import logging
2+
from typing import List
3+
14
import pandas as pd
25
from zenml import step
3-
from typing import List
4-
import logging
56

67
# Set up logger
78
logger = logging.getLogger(__name__)

bank_subscription_prediction/steps/data_splitter.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,9 @@
1+
import logging
2+
from typing import Annotated, Tuple
3+
14
import pandas as pd
2-
from zenml import step
35
from sklearn.model_selection import train_test_split
4-
from typing import Tuple, Annotated
5-
import logging
6+
from zenml import step
67

78
# Set up logger
89
logger = logging.getLogger(__name__)

bank_subscription_prediction/steps/model_evaluator.py

Lines changed: 7 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -1,26 +1,20 @@
1+
from typing import Dict, List, Tuple
2+
3+
import numpy as np
14
import pandas as pd
5+
import plotly.express as px
6+
import plotly.graph_objects as go
27
import xgboost as xgb
38
from sklearn.feature_selection import SelectFromModel
49
from sklearn.metrics import (
510
accuracy_score,
6-
classification_report,
11+
auc,
712
confusion_matrix,
8-
roc_curve,
913
precision_recall_curve,
10-
auc,
11-
roc_auc_score,
14+
roc_curve,
1215
)
13-
import matplotlib.pyplot as plt
14-
import plotly.express as px
15-
import plotly.graph_objects as go
16-
from plotly.subplots import make_subplots
17-
import numpy as np
18-
import json
19-
from typing import Dict, Any, List, Tuple
2016
from typing_extensions import Annotated
21-
2217
from zenml import step
23-
from zenml.materializers.built_in_materializer import BuiltInMaterializer
2418
from zenml.types import HTMLString
2519

2620

0 commit comments

Comments
 (0)