Skip to content

Commit 5c87056

Browse files
authored
Merge pull request #419 from middlewarehq/dependabot/pip/backend/pip-ec4054ff4d
Bump the pip group across 1 directory with 4 updates
2 parents 8c314eb + ebeb22e commit 5c87056

30 files changed

+231
-218
lines changed

.github/workflows/black.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ jobs:
2727
python-version: "3.11.6"
2828

2929
- name: Install Black
30-
run: python -m pip install black==22.3.0
30+
run: python -m pip install black==24.3.0
3131

3232
- name: Run Black Check
3333
run: black . --check

backend/analytics_server/mhq/api/deployment_analytics.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -127,9 +127,9 @@ def get_prs_included_in_deployment(deployment_id: str):
127127

128128
repo: OrgRepo = pr_analytics_service.get_repo_by_id(deployment.repo_id)
129129

130-
prs: List[
131-
PullRequest
132-
] = deployments_service.get_pull_requests_related_to_deployment(deployment)
130+
prs: List[PullRequest] = (
131+
deployments_service.get_pull_requests_related_to_deployment(deployment)
132+
)
133133
repo_id_map = {repo.id: repo}
134134

135135
return get_non_paginated_pr_response(
@@ -204,10 +204,10 @@ def get_team_deployment_frequency_trends(
204204

205205
deployments_analytics_service = get_deployment_analytics_service()
206206

207-
week_to_deployments_count_map: Dict[
208-
datetime, int
209-
] = deployments_analytics_service.get_weekly_deployment_frequency_trends(
210-
team_id, interval, pr_filter, workflow_filter
207+
week_to_deployments_count_map: Dict[datetime, int] = (
208+
deployments_analytics_service.get_weekly_deployment_frequency_trends(
209+
team_id, interval, pr_filter, workflow_filter
210+
)
211211
)
212212

213213
return {

backend/analytics_server/mhq/api/incidents.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -92,9 +92,9 @@ def get_deployments_with_related_incidents(
9292

9393
incidents: List[Incident] = incident_service.get_team_incidents(team_id, interval)
9494

95-
deployment_incidents_map: Dict[
96-
Deployment, List[Incident]
97-
] = incident_service.get_deployment_incidents_map(deployments, incidents)
95+
deployment_incidents_map: Dict[Deployment, List[Incident]] = (
96+
incident_service.get_deployment_incidents_map(deployments, incidents)
97+
)
9898

9999
return list(
100100
map(
@@ -238,10 +238,10 @@ def get_team_cfr_trends(
238238

239239
incidents: List[Incident] = incident_service.get_team_incidents(team_id, interval)
240240

241-
team_weekly_change_failure_rate: Dict[
242-
datetime, ChangeFailureRateMetrics
243-
] = incident_service.get_weekly_change_failure_rate(
244-
interval, deployments, incidents
241+
team_weekly_change_failure_rate: Dict[datetime, ChangeFailureRateMetrics] = (
242+
incident_service.get_weekly_change_failure_rate(
243+
interval, deployments, incidents
244+
)
245245
)
246246

247247
return {

backend/analytics_server/mhq/api/pull_requests.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -159,9 +159,9 @@ def get_team_lead_time_trends(
159159

160160
lead_time_service = get_lead_time_service()
161161

162-
weekly_lead_time_metrics_avg_map: Dict[
163-
datetime, LeadTimeMetrics
164-
] = lead_time_service.get_team_lead_time_metrics_trends(team, interval, pr_filter)
162+
weekly_lead_time_metrics_avg_map: Dict[datetime, LeadTimeMetrics] = (
163+
lead_time_service.get_team_lead_time_metrics_trends(team, interval, pr_filter)
164+
)
165165

166166
return {
167167
week.isoformat(): adapt_lead_time_metrics(average_lead_time_metrics)

backend/analytics_server/mhq/api/resources/incident_resources.py

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -22,12 +22,14 @@ def adapt_incident(
2222
"provider": incident.provider,
2323
"status": incident.status,
2424
"creation_date": incident.creation_date.isoformat(),
25-
"resolved_date": incident.resolved_date.isoformat()
26-
if incident.resolved_date
27-
else None,
28-
"acknowledged_date": incident.acknowledged_date.isoformat()
29-
if incident.acknowledged_date
30-
else None,
25+
"resolved_date": (
26+
incident.resolved_date.isoformat() if incident.resolved_date else None
27+
),
28+
"acknowledged_date": (
29+
incident.acknowledged_date.isoformat()
30+
if incident.acknowledged_date
31+
else None
32+
),
3133
"assigned_to": adapt_user_info(incident.assigned_to, username_user_map),
3234
"assignees": list(
3335
map(

backend/analytics_server/mhq/exapi/git_incidents.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -40,10 +40,10 @@ def get_org_repo(self, repo_id: str):
4040
def get_repo_revert_prs_in_interval(
4141
self, repo_id: str, from_time: datetime, to_time: datetime
4242
) -> List[RevertPRMap]:
43-
revert_pr_mappings: List[
44-
PullRequestRevertPRMapping
45-
] = self.code_repo_service.get_repo_revert_prs_mappings_updated_in_interval(
46-
repo_id, from_time, to_time
43+
revert_pr_mappings: List[PullRequestRevertPRMapping] = (
44+
self.code_repo_service.get_repo_revert_prs_mappings_updated_in_interval(
45+
repo_id, from_time, to_time
46+
)
4747
)
4848

4949
revert_pr_ids = [str(pr.pr_id) for pr in revert_pr_mappings]

backend/analytics_server/mhq/service/code/integration.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -13,10 +13,10 @@ def __init__(self, core_repo_service: CoreRepoService):
1313
self.core_repo_service = core_repo_service
1414

1515
def get_org_providers(self, org_id: str) -> List[str]:
16-
integrations: List[
17-
Integration
18-
] = self.core_repo_service.get_org_integrations_for_names(
19-
org_id, CODE_INTEGRATION_BUCKET
16+
integrations: List[Integration] = (
17+
self.core_repo_service.get_org_integrations_for_names(
18+
org_id, CODE_INTEGRATION_BUCKET
19+
)
2020
)
2121
if not integrations:
2222
return []

backend/analytics_server/mhq/service/code/lead_time.py

Lines changed: 16 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -54,15 +54,15 @@ def get_team_lead_time_metrics_trends(
5454
set(self._get_team_repos_lead_time_metrics(team_repos, interval, pr_filter))
5555
)
5656

57-
weekly_lead_time_metrics_map: Dict[
58-
datetime, List[LeadTimeMetrics]
59-
] = generate_expanded_buckets(
60-
lead_time_metrics, interval, "merged_at", "weekly"
57+
weekly_lead_time_metrics_map: Dict[datetime, List[LeadTimeMetrics]] = (
58+
generate_expanded_buckets(
59+
lead_time_metrics, interval, "merged_at", "weekly"
60+
)
6161
)
6262

63-
weekly_lead_time_metrics_avg_map: Dict[
64-
datetime, LeadTimeMetrics
65-
] = self.get_avg_lead_time_metrics_from_map(weekly_lead_time_metrics_map)
63+
weekly_lead_time_metrics_avg_map: Dict[datetime, LeadTimeMetrics] = (
64+
self.get_avg_lead_time_metrics_from_map(weekly_lead_time_metrics_map)
65+
)
6666

6767
weekly_lead_time_metrics_avg_map = fill_missing_week_buckets(
6868
weekly_lead_time_metrics_avg_map, interval, LeadTimeMetrics
@@ -176,10 +176,10 @@ def _get_lead_time_prs_for_repos_using_workflow_deployments(
176176
pr_filter: PRFilter = None,
177177
) -> List[PullRequest]:
178178

179-
team_repos_with_workflow_deployments_configured: List[
180-
TeamRepos
181-
] = self._deployments_service.get_filtered_team_repos_with_workflow_configured_deployments(
182-
team_repos
179+
team_repos_with_workflow_deployments_configured: List[TeamRepos] = (
180+
self._deployments_service.get_filtered_team_repos_with_workflow_configured_deployments(
181+
team_repos
182+
)
183183
)
184184

185185
repo_ids = [
@@ -214,9 +214,11 @@ def _get_lead_time_prs_for_repos_using_pr_deployments(
214214

215215
def _get_lead_time_metrics_for_pr(self, pr: PullRequest) -> LeadTimeMetrics:
216216
return LeadTimeMetrics(
217-
first_commit_to_open=pr.first_commit_to_open
218-
if pr.first_commit_to_open is not None and pr.first_commit_to_open > 0
219-
else 0,
217+
first_commit_to_open=(
218+
pr.first_commit_to_open
219+
if pr.first_commit_to_open is not None and pr.first_commit_to_open > 0
220+
else 0
221+
),
220222
first_response_time=pr.first_response_time if pr.first_response_time else 0,
221223
rework_time=pr.rework_time if pr.rework_time else 0,
222224
merge_time=pr.merge_time if pr.merge_time else 0,

backend/analytics_server/mhq/service/code/repository_service.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -194,9 +194,9 @@ def _adapt_org_incident_service(
194194
name=org_repo.name,
195195
key=str(org_repo.id),
196196
meta={},
197-
created_at=org_incident_service.created_at
198-
if org_incident_service
199-
else time_now(),
197+
created_at=(
198+
org_incident_service.created_at if org_incident_service else time_now()
199+
),
200200
updated_at=time_now(),
201201
source_type=IncidentSource.GIT_REPO,
202202
)

backend/analytics_server/mhq/service/code/sync/etl_code_analytics.py

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -92,9 +92,11 @@ def get_pr_performance(pr: PullRequest, pr_events: [PullRequestEvent]):
9292
cycle_time = cycle_time.total_seconds()
9393

9494
return PRPerformance(
95-
first_review_time=(first_review.created_at - pr.created_at).total_seconds()
96-
if first_review
97-
else -1,
95+
first_review_time=(
96+
(first_review.created_at - pr.created_at).total_seconds()
97+
if first_review
98+
else -1
99+
),
98100
rework_time=rework_time,
99101
merge_time=merge_time,
100102
cycle_time=cycle_time if pr.state == PullRequestState.MERGED else -1,

0 commit comments

Comments
 (0)