Skip to content

performance(check-agent): test to reduce buffer size numbers#665

Closed
GustavoCaso wants to merge 6 commits intomainfrom
gustavo.caso/reduce-checks-agent-buffer-size
Closed

performance(check-agent): test to reduce buffer size numbers#665
GustavoCaso wants to merge 6 commits intomainfrom
gustavo.caso/reduce-checks-agent-buffer-size

Conversation

@GustavoCaso
Copy link
Contributor

@GustavoCaso GustavoCaso commented May 6, 2025

Summary

Reduce buffer size to see if the RSS footprint is affected by this change

Change Type

  • Bug fix
  • New feature
  • Non-functional (chore, refactoring, docs)
  • Performance

How did you test this PR?

References

@GustavoCaso GustavoCaso changed the title test reduce numbers for checks agent buffer size (check-agent): test to reduce buffer size numbers May 6, 2025
@pr-commenter
Copy link

pr-commenter bot commented May 6, 2025

Regression Detector (DogStatsD)

Regression Detector Results

Run ID: 6c68e8e7-3381-41dd-a31d-3b6b6566b308

Baseline: 7.65.2
Comparison: 7.65.2

Optimization Goals: ✅ No significant changes detected

Fine details of change detection per experiment

perf experiment goal Δ mean % Δ mean % CI trials links
dsd_uds_500mb_3k_contexts ingress throughput +0.47 [+0.39, +0.55] 1
dsd_uds_100mb_3k_contexts_distributions_only memory utilization +0.37 [+0.16, +0.58] 1
quality_gates_idle_rss memory utilization +0.19 [+0.09, +0.29] 1
dsd_uds_1mb_3k_contexts ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_1mb_50k_contexts ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_512kb_3k_contexts ingress throughput +0.00 [-0.01, +0.01] 1
dsd_uds_1mb_50k_contexts_memlimit ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_1mb_3k_contexts_dualship ingress throughput -0.00 [-0.00, +0.00] 1
dsd_uds_100mb_3k_contexts ingress throughput -0.00 [-0.10, +0.10] 1
dsd_uds_40mb_12k_contexts_40_senders ingress throughput -0.00 [-0.00, +0.00] 1
dsd_uds_100mb_250k_contexts ingress throughput -0.00 [-0.10, +0.10] 1
dsd_uds_10mb_3k_contexts ingress throughput -0.01 [-0.08, +0.07] 1

Bounds Checks: ❌ Failed

perf experiment bounds_check_name replicates_passed links
quality_gates_idle_rss memory_usage 0/10

Explanation

Confidence level: 90.00%
Effect size tolerance: |Δ mean %| ≥ 5.00%

Performance changes are noted in the perf column of each table:

  • ✅ = significantly better comparison variant performance
  • ❌ = significantly worse comparison variant performance
  • ➖ = no significant change in performance

A regression test is an A/B test of target performance in a repeatable rig, where "performance" is measured as "comparison variant minus baseline variant" for an optimization goal (e.g., ingress throughput). Due to intrinsic variability in measuring that goal, we can only estimate its mean value for each experiment; we report uncertainty in that value as a 90.00% confidence interval denoted "Δ mean % CI".

For each experiment, we decide whether a change in performance is a "regression" -- a change worth investigating further -- if all of the following criteria are true:

  1. Its estimated |Δ mean %| ≥ 5.00%, indicating the change is big enough to merit a closer look.

  2. Its 90.00% confidence interval "Δ mean % CI" does not contain zero, indicating that if our statistical model is accurate, there is at least a 90.00% chance there is a difference in performance between baseline and comparison variants.

  3. Its configuration does not mark it "erratic".

@pr-commenter
Copy link

pr-commenter bot commented May 6, 2025

Regression Detector (Checks Agent Go)

Regression Detector Results

Run ID: acda33ec-31c1-463f-a7ea-d7685d9848f7

Baseline: f61d1f4e054b884cb1894254ab2714b84b4684cb
Comparison: f61d1f4e054b884cb1894254ab2714b84b4684cb
Diff

Optimization Goals: ✅ No significant changes detected

Fine details of change detection per experiment

perf experiment goal Δ mean % Δ mean % CI trials links
quality_gates_rss memory utilization +0.02 [-0.04, +0.08] 1
quality_gates_idle_rss memory utilization -0.16 [-0.21, -0.11] 1

Explanation

Confidence level: 90.00%
Effect size tolerance: |Δ mean %| ≥ 5.00%

Performance changes are noted in the perf column of each table:

  • ✅ = significantly better comparison variant performance
  • ❌ = significantly worse comparison variant performance
  • ➖ = no significant change in performance

A regression test is an A/B test of target performance in a repeatable rig, where "performance" is measured as "comparison variant minus baseline variant" for an optimization goal (e.g., ingress throughput). Due to intrinsic variability in measuring that goal, we can only estimate its mean value for each experiment; we report uncertainty in that value as a 90.00% confidence interval denoted "Δ mean % CI".

For each experiment, we decide whether a change in performance is a "regression" -- a change worth investigating further -- if all of the following criteria are true:

  1. Its estimated |Δ mean %| ≥ 5.00%, indicating the change is big enough to merit a closer look.

  2. Its 90.00% confidence interval "Δ mean % CI" does not contain zero, indicating that if our statistical model is accurate, there is at least a 90.00% chance there is a difference in performance between baseline and comparison variants.

  3. Its configuration does not mark it "erratic".

@pr-commenter
Copy link

pr-commenter bot commented May 6, 2025

Regression Detector (Saluki)

Regression Detector Results

Run ID: 9af277fc-11c8-487a-a0f7-691298ec19aa

Baseline: 7b38aae
Comparison: c86cb28
Diff

Optimization Goals: ✅ No significant changes detected

Fine details of change detection per experiment

perf experiment goal Δ mean % Δ mean % CI trials links
quality_gates_idle_rss memory utilization +0.78 [+0.73, +0.83] 1
dsd_uds_40mb_12k_contexts_40_senders ingress throughput +0.32 [+0.20, +0.45] 1
dsd_uds_10mb_3k_contexts ingress throughput +0.01 [-0.02, +0.04] 1
dsd_uds_512kb_3k_contexts ingress throughput +0.01 [-0.01, +0.02] 1
dsd_uds_1mb_50k_contexts_memlimit ingress throughput +0.01 [-0.00, +0.01] 1
dsd_uds_50mb_10k_contexts_no_inlining ingress throughput -0.00 [-0.10, +0.10] 1
dsd_uds_50mb_10k_contexts_no_inlining_no_allocs ingress throughput -0.00 [-0.11, +0.11] 1
dsd_uds_1mb_50k_contexts ingress throughput -0.01 [-0.02, +0.00] 1
dsd_uds_1mb_3k_contexts_dualship ingress throughput -0.01 [-0.02, +0.01] 1
dsd_uds_100mb_3k_contexts ingress throughput -0.01 [-0.08, +0.06] 1
dsd_uds_1mb_3k_contexts ingress throughput -0.01 [-0.02, +0.00] 1
dsd_uds_500mb_3k_contexts ingress throughput -0.31 [-0.43, -0.19] 1
dsd_uds_100mb_250k_contexts ingress throughput -0.56 [-0.74, -0.38] 1
dsd_uds_100mb_3k_contexts_distributions_only memory utilization -0.84 [-0.99, -0.68] 1

Bounds Checks: ✅ Passed

perf experiment bounds_check_name replicates_passed links
quality_gates_idle_rss memory_usage 10/10

Explanation

Confidence level: 90.00%
Effect size tolerance: |Δ mean %| ≥ 5.00%

Performance changes are noted in the perf column of each table:

  • ✅ = significantly better comparison variant performance
  • ❌ = significantly worse comparison variant performance
  • ➖ = no significant change in performance

A regression test is an A/B test of target performance in a repeatable rig, where "performance" is measured as "comparison variant minus baseline variant" for an optimization goal (e.g., ingress throughput). Due to intrinsic variability in measuring that goal, we can only estimate its mean value for each experiment; we report uncertainty in that value as a 90.00% confidence interval denoted "Δ mean % CI".

For each experiment, we decide whether a change in performance is a "regression" -- a change worth investigating further -- if all of the following criteria are true:

  1. Its estimated |Δ mean %| ≥ 5.00%, indicating the change is big enough to merit a closer look.

  2. Its 90.00% confidence interval "Δ mean % CI" does not contain zero, indicating that if our statistical model is accurate, there is at least a 90.00% chance there is a difference in performance between baseline and comparison variants.

  3. Its configuration does not mark it "erratic".

@pr-commenter
Copy link

pr-commenter bot commented May 6, 2025

Regression Detector (Checks Agent)

Regression Detector Results

Run ID: 4221abd2-6a51-495a-a515-2e8d29eac0ac

Baseline: 7b38aae
Comparison: c86cb28
Diff

Optimization Goals: ✅ No significant changes detected

Fine details of change detection per experiment

perf experiment goal Δ mean % Δ mean % CI trials links
quality_gates_rss memory utilization -0.51 [-0.52, -0.49] 1
quality_gates_idle_rss memory utilization -0.86 [-0.87, -0.84] 1

Bounds Checks: ❌ Failed

perf experiment bounds_check_name replicates_passed links
quality_gates_rss memory_usage 0/10
quality_gates_idle_rss memory_usage 10/10

Explanation

Confidence level: 90.00%
Effect size tolerance: |Δ mean %| ≥ 5.00%

Performance changes are noted in the perf column of each table:

  • ✅ = significantly better comparison variant performance
  • ❌ = significantly worse comparison variant performance
  • ➖ = no significant change in performance

A regression test is an A/B test of target performance in a repeatable rig, where "performance" is measured as "comparison variant minus baseline variant" for an optimization goal (e.g., ingress throughput). Due to intrinsic variability in measuring that goal, we can only estimate its mean value for each experiment; we report uncertainty in that value as a 90.00% confidence interval denoted "Δ mean % CI".

For each experiment, we decide whether a change in performance is a "regression" -- a change worth investigating further -- if all of the following criteria are true:

  1. Its estimated |Δ mean %| ≥ 5.00%, indicating the change is big enough to merit a closer look.

  2. Its 90.00% confidence interval "Δ mean % CI" does not contain zero, indicating that if our statistical model is accurate, there is at least a 90.00% chance there is a difference in performance between baseline and comparison variants.

  3. Its configuration does not mark it "erratic".

@pr-commenter
Copy link

pr-commenter bot commented May 6, 2025

Regression Detector Links

ADP Experiment Result Links

experiment link(s)
dsd_uds_100mb_250k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_100mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_100mb_3k_contexts_distributions_only [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_10mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_3k_contexts_dualship [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_50k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_50k_contexts_memlimit [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_40mb_12k_contexts_40_senders [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_500mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_512kb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
quality_gates_idle_rss [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_50mb_10k_contexts_no_inlining (ADP only) [Profiling (ADP)] [SMP Dashboard]
dsd_uds_50mb_10k_contexts_no_inlining_no_allocs (ADP only) [Profiling (ADP)] [SMP Dashboard]

Checks Agent Experiment Result Links

experiment link(s)
quality_gates_idle_rss [Profiling] [SMP Dashboard]
quality_gates_rss [Profiling] [SMP Dashboard]

@GustavoCaso GustavoCaso force-pushed the gustavo.caso/reduce-checks-agent-buffer-size branch from a4ec0b8 to 3e6daac Compare May 7, 2025 07:56
Copy link
Member

tobz commented May 7, 2025

@GustavoCaso GustavoCaso changed the title (check-agent): test to reduce buffer size numbers performance(check-agent): test to reduce buffer size numbers Jun 10, 2025
@GustavoCaso GustavoCaso marked this pull request as ready for review June 10, 2025 08:15
@GustavoCaso GustavoCaso requested a review from a team as a code owner June 10, 2025 08:15
@tobz tobz deleted the gustavo.caso/reduce-checks-agent-buffer-size branch September 24, 2025 18:21
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants