-
Notifications
You must be signed in to change notification settings - Fork 25
ci: add scheduled benchmark runs (#1639) #1645
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you please rebase this PR on top of #1642 (or whatever PR it's dependent on)? I'm not what's needed to be reviewed vs code that's being reviewed in a different PR.
| summary-always: true | ||
| auto-push: true | ||
| fail-on-alert: true |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
q: what's the rationale for setting summary-always and fail-on-alert to true?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Both are about visibility and catching regressions early. summary-always ensures results are always visible in the job summary. fail-on-alert ensures we don't silently regress - if performance drops significantly, the workflow fails rather than quietly recording bad data.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
if performance drops significantly
Hmm what does it mean for performance to drop significantly?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good question, I've added explicit alert-threshold: "150%" to define what "significant" means - the workflow will fail if performance degrades by more than 50% compared to the previous run.
fail-on-alert signals that the workflow failed specifically because of a performance regression (vs. an infrastructure issue), and alert-threshold defines the sensitivity. We can tune this based on observed variance once we have more data points.
rkuris
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nit: Overall, this seems like a lot of somewhat-fragile code to parse json instead of just specifying the parameters in the normal github style.
I think we can save a ton of money by picking a much smaller instance size. I think i4i.xlarge will work at roughly 1/8 the cost.
Implements workflow to trigger C-Chain reexecution benchmarks in AvalancheGo and track Firewood performance over time. Supports task-based and custom parameter modes. Results stored in GitHub Pages via github-action-benchmark. # Conflicts: # .github/workflows/track-performance.yml
- Refactor `bench-cchain` to trigger track-performance.yml instead of
directly calling AvalancheGo's workflow
- Add input validation and helpful error messages for test/custom params
- Use commit SHA instead of branch name for reproducibility
- Fix AvalancheGo workflow input: use `with-dependencies` format
("firewood=abc,libevm=xyz") instead of separate firewood-ref/libevm-ref
- Remove status-cchain, list-cchain, help-cchain (use GitHub UI instead)
- Remove debug logging from track-performance.yml
Remove local developer tooling (justfile recipe, flake.nix, METRICS.md) to reduce PR scope. These will be submitted in a follow-up PR after the CI workflow changes are merged.
Daily (40M→41M) and weekly (50M→60M) benchmarks with JSON-based config and matrix strategy for cleaner workflow management.
2af94f6 to
6ee9ccc
Compare
…coded branch name used for temp. testing
- Replace `avago-runner-i4i-4xlarge-local-ssd` with `avago-runner-i4i-2xlarge-local-ssd` for daily and weekly benchmarks. - Add new input options for `firewood` and `libevm` commits/branches/tags. - Extend runner choices with additional configurations in the workflow.
- Remove `.github/benchmark-schedules.json` in favor of inline configurations. - Replace matrix-based strategy with direct output handling for cleaner and more explicit workflow logic. - Preserve manual and scheduled benchmark support with optimized input handling.
Updated was trying to stay consistent with AvalancheGo's JSON config approach but I agree with you inline with native GitHub outputs feels more natural here. No strong need for consistency on this at this point in time. |
…ale workspace issues
|
Test case manual dispatch with custom dependencies Command: Run: https://github.com/ava-labs/firewood/actions/runs/21489879261
Graph: https://ava-labs.github.io/firewood/dev/bench/es-scheduled-perf-tracking/ |
- Add workflow_dispatch trigger to gh-pages.yaml for manual runs - Trigger gh-pages deployment from track-performance.yml after benchmark results are pushed to benchmark-data branch This ensures benchmark results are immediately visible on GitHub Pages without waiting for a push to main.
Avoids conflict with daily run (which includes Friday) and ensures 10M block benchmark results are available early in the week.
Why this should be merged
This patch is closing #1639 it adds automated benchmark triggers to catch performance regressions without manual intervention.
Note: should be merged after #1493
How this works
Manual dispatch still works via workflow inputs.
How this was tested