Skip to content

Commit 04696a5

Browse files
authored
run formatting (#186)
1 parent feef5d3 commit 04696a5

File tree

24 files changed

+153
-122
lines changed

24 files changed

+153
-122
lines changed

databricks-production-qa-demo/steps/deployment/deployment_deploy.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -31,10 +31,14 @@
3131

3232

3333
@step(enable_cache=False)
34-
def deployment_deploy() -> Annotated[
35-
Optional[DatabricksDeploymentService],
36-
ArtifactConfig(name="databricks_deployment", is_deployment_artifact=True),
37-
]:
34+
def deployment_deploy() -> (
35+
Annotated[
36+
Optional[DatabricksDeploymentService],
37+
ArtifactConfig(
38+
name="databricks_deployment", is_deployment_artifact=True
39+
),
40+
]
41+
):
3842
"""Predictions step.
3943
4044
This is an example of a predictions step that takes the data in and returns

end-to-end-computer-vision/steps/export_label_studio.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,6 @@
2828
logger = get_logger(__name__)
2929

3030

31-
3231
@step(
3332
output_materializers={
3433
LABELED_DATASET_NAME: LabelStudioAnnotationMaterializer

eurorate-predictor/pipelines/training.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,9 @@
2424

2525

2626
@pipeline
27-
def ecb_predictor_model_training_pipeline(augmented_dataset_id, mode: str = "develop"):
27+
def ecb_predictor_model_training_pipeline(
28+
augmented_dataset_id, mode: str = "develop"
29+
):
2830
"""A pipeline to train an XGBoost model and promote it.
2931
3032
Args:

eurorate-predictor/run.py

Lines changed: 9 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -101,7 +101,9 @@ def main(
101101
pipeline_args["config_path"] = os.path.join(
102102
config_folder, f"etl_{mode}.yaml"
103103
)
104-
ecb_predictor_etl_pipeline.with_options(**pipeline_args)(**run_args_etl)
104+
ecb_predictor_etl_pipeline.with_options(**pipeline_args)(
105+
**run_args_etl
106+
)
105107
logger.info("ETL pipeline finished successfully!\n")
106108

107109
# Execute Feature Engineering Pipeline
@@ -126,9 +128,9 @@ def main(
126128
pipeline_args["config_path"] = os.path.join(
127129
config_folder, f"feature_engineering_{mode}.yaml"
128130
)
129-
ecb_predictor_feature_engineering_pipeline.with_options(**pipeline_args)(
130-
**run_args_feature
131-
)
131+
ecb_predictor_feature_engineering_pipeline.with_options(
132+
**pipeline_args
133+
)(**run_args_feature)
132134
logger.info("Feature Engineering pipeline finished successfully!\n")
133135

134136
# Execute Model Training Pipeline
@@ -153,7 +155,9 @@ def main(
153155
pipeline_args["config_path"] = os.path.join(
154156
config_folder, f"training_{mode}.yaml"
155157
)
156-
ecb_predictor_model_training_pipeline.with_options(**pipeline_args)(**run_args_train)
158+
ecb_predictor_model_training_pipeline.with_options(**pipeline_args)(
159+
**run_args_train
160+
)
157161
logger.info("Model Training pipeline finished successfully!\n")
158162

159163

gamesense/steps/finetune.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,12 +28,11 @@
2828
from utils.loaders import load_base_model
2929
from utils.tokenizer import load_tokenizer
3030
from zenml import ArtifactConfig, step
31+
from zenml.client import Client
3132
from zenml.enums import ArtifactType
3233
from zenml.logger import get_logger
3334
from zenml.materializers import BuiltInMaterializer
3435
from zenml.utils.cuda_utils import cleanup_gpu_memory
35-
from zenml.client import Client
36-
3736

3837
logger = get_logger(__name__)
3938

huggingface-sagemaker/steps/deploying/huggingface_deployment.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -47,9 +47,7 @@ def deploy_to_huggingface(
4747
save_model_to_deploy.entrypoint()
4848

4949
logger.info("Model saved locally. Pushing to HuggingFace...")
50-
assert secret, (
51-
"No secret found with name 'huggingface_creds'. Please create one with your `token`."
52-
)
50+
assert secret, "No secret found with name 'huggingface_creds'. Please create one with your `token`."
5351

5452
token = secret.secret_values["token"]
5553
api = HfApi(token=token)

huggingface-sagemaker/steps/promotion/promote_get_metrics.py

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -27,10 +27,12 @@
2727

2828

2929
@step
30-
def promote_get_metrics() -> Tuple[
31-
Annotated[Dict[str, Any], "latest_metrics"],
32-
Annotated[Dict[str, Any], "current_metrics"],
33-
]:
30+
def promote_get_metrics() -> (
31+
Tuple[
32+
Annotated[Dict[str, Any], "latest_metrics"],
33+
Annotated[Dict[str, Any], "current_metrics"],
34+
]
35+
):
3436
"""Get metrics for comparison for promoting a model.
3537
3638
This is an example of a metric retrieval step. It is used to retrieve

llm-complete-guide/pipelines/llm_eval.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717
from typing import Optional
1818

1919
import click
20-
from steps.create_prompt import PROMPT, create_prompt
20+
from steps.create_prompt import PROMPT
2121
from steps.eval_e2e import e2e_evaluation, e2e_evaluation_llm_judged
2222
from steps.eval_retrieval import (
2323
retrieval_evaluation_full,

llm-complete-guide/steps/create_prompt.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@
2424
answers. \
2525
"""
2626

27+
2728
@step
2829
def create_prompt() -> str:
2930
"""Create a prompt for the RAG pipeline."""

llm-complete-guide/steps/eval_retrieval.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -275,9 +275,9 @@ def perform_small_retrieval_evaluation(use_reranking: bool) -> float:
275275

276276

277277
@step
278-
def retrieval_evaluation_small() -> Annotated[
279-
float, "small_failure_rate_retrieval"
280-
]:
278+
def retrieval_evaluation_small() -> (
279+
Annotated[float, "small_failure_rate_retrieval"]
280+
):
281281
"""Executes the retrieval evaluation step without reranking.
282282
283283
Returns:
@@ -287,9 +287,9 @@ def retrieval_evaluation_small() -> Annotated[
287287

288288

289289
@step
290-
def retrieval_evaluation_small_with_reranking() -> Annotated[
291-
float, "small_failure_rate_retrieval_reranking"
292-
]:
290+
def retrieval_evaluation_small_with_reranking() -> (
291+
Annotated[float, "small_failure_rate_retrieval_reranking"]
292+
):
293293
"""Executes the retrieval evaluation step with reranking.
294294
295295
Returns:

0 commit comments

Comments
 (0)