Skip to content

[integ-tests-framework] Add OS comparison sheet to performance summary#7248

Merged
hanwen-cluster merged 2 commits intoaws:developfrom
hanwen-cluster:developfeb16
Mar 2, 2026
Merged

[integ-tests-framework] Add OS comparison sheet to performance summary#7248
hanwen-cluster merged 2 commits intoaws:developfrom
hanwen-cluster:developfeb16

Conversation

@hanwen-cluster
Copy link
Contributor

Description of changes

Add an "OS Comparison" sheet to performance_summary.xlsx that shows average performance (last 30 days) per OS for each benchmark, making it easy to spot OS-specific anomalies. Features:

  • One row per benchmark, one column per OS, showing 30-day average values
  • Green cell = best performing OS, red cell = worst performing OS (direction-aware for bibw vs latency)
  • Spread column = (worst - best) / best, showing how much OS choice matters for each benchmark
  • Spread coloring: yellow >25%, red >50%
  • Existing trend analysis moved to "Trend" sheet in the same file

Tests

  • Tested the summary is correctly generated

References

  • Link to impacted open issues.
  • Link to related PRs in other packages (i.e. cookbook, node).
  • Link to documentation useful to understand the changes.

Checklist

  • Make sure you are pointing to the right branch.
  • If you're creating a patch for a branch other than develop add the branch name as prefix in the PR title (e.g. [release-3.6]).
  • Check all commits' messages are clear, describing what and why vs how.
  • Make sure to have added unit tests or integration tests to cover the new/modified code.
  • Check if documentation is impacted by this change.

Please review the guidelines for contributing and Pull Request Instructions.

By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.

Add an "OS Comparison" sheet to performance_summary.xlsx that shows average performance (last 30 days) per OS for each benchmark, making it easy to spot OS-specific anomalies.
Features:
* One row per benchmark, one column per OS, showing 30-day average values
* Green cell = best performing OS, red cell = worst performing OS (direction-aware for bibw vs latency)
* Spread column = (worst - best) / best, showing how much OS choice matters for each benchmark
* Spread coloring: yellow >25%, red >50%
* Existing trend analysis moved to "Trend" sheet in the same file
@hanwen-cluster hanwen-cluster requested review from a team as code owners February 25, 2026 22:07
@hanwen-cluster hanwen-cluster added the skip-changelog-update Disables the check that enforces changelog updates in PRs label Feb 25, 2026
if not os_averages:
continue

# Calculate spread
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

[NON_blocking] I dont have a problem keeping the Spread calculation in the sheet, but what is the advantage or gain we see in comparing a specific benchmarks across OS? Do we use this for a comparision of 2 OSSes packages/kernels, etc so that if find out what is different in the "best OS" compared to the "worst OS"?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is helpful for example to discover aws/aws-parallelcluster-cookbook#3130

@hanwen-cluster hanwen-cluster enabled auto-merge (rebase) March 2, 2026 19:27
@hanwen-cluster hanwen-cluster merged commit aa5440d into aws:develop Mar 2, 2026
24 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

skip-changelog-update Disables the check that enforces changelog updates in PRs

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants