-
-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
General Retrospective for January 2025 Releases #64
Comments
Release pipelines which have some errors in the downstream jobs are not able to generate the release summary report in TRSS. SL/Jan15: think this is related to the amount of failures to the point where it essentially exceeds the character limit in the release report. Pipelines with a some errors can be generated, pipelines with a massive amount of information to 'share' exceed the limit. aqa-test-tools/issues/xxxx to print out a msg if limit is hit (so the user knows no report will be generated). |
I have manually increased the TIME_LIMIT on https://ci.adoptium.net/job/Test_openjdk23_hs_extended.openjdk_riscv64_linux/ as it appears to be hitting its 25 hour limit and aborting before completion. FYI @Haroon-Khel We will need to investigate why its timing out (my guess is that certain testcase failures hanging and running long until each one hits its timeout). If there is time, we should figure that out during the dry run assessment, but if not a longer time limit will hopefully allow the jobs to finish without aborting even with the timeouts. |
https://adoptium.slack.com/archives/C09NW3L2J/p1737466419332739
Steps to resolve this from Andrew:
|
Some test jobs timeout during this release and dry run even with timeout=25 hours. The reason is TRSS unavailable and based on assumptive test times ( very random and small number) testlist number is set as 1. So not parallel at all for release jdk21,17,11,8. Only jdk23 tests are parallel. This is why for this release even for some primary platforms tests run slowly. The timeout caused build failed and no test results are archived, we have to rerun the job to get the test results. For example:
It may be better to fall back to parallel by a pre-defined numberofnode if this happens ( even though rarely happens). |
Should "Re-run in Grinder" links always set For example at the time of originally posting this comment it's only showing two jobs in the main display: despite there being other jobs prior to that still running which were initiated with the re-run links which have |
Should we cherry pick cacerts updates that occur in the master branch between the branching for the dry-run and the final GA builds?
|
Summary
A retrospective for all efforts surrounding the titular releases.
All community members are welcome to contribute to the agenda via comments below.
This will be a virtual meeting after the release, with at least a week of notice in the #release Slack channel.
On the day of the meeting we'll review the agenda and add a list of actions at the end.
Invited: Everyone.
Time, Date, and URL
Time:
Date:
URL:
Details
Retrospective Owner Tasks (in order):
TLDR
Add proposed agenda items as comments below.
The text was updated successfully, but these errors were encountered: