Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cirrus: reduce task timeout #5994

Merged
merged 1 commit into from
Feb 19, 2025

Conversation

Luap99
Copy link
Member

@Luap99 Luap99 commented Feb 19, 2025

With all the recent speed-ups here the timeout is way to high, all tasks should complete in under 30 mins generally. The smoke test in under 10min as it does not do much.

In particular I noticed at least two separate rootless integration tests time out after 120min[1,2], obviously the tests do not take that long and they are hanging somehwere instead. With a lower timeout we do not waste so much time when this happens.

[1] https://cirrus-ci.com/task/4733420225429504
[2] https://cirrus-ci.com/task/5597909967699968

What type of PR is this?

/kind api-change
/kind bug
/kind cleanup
/kind deprecation
/kind design
/kind documentation
/kind failing-test
/kind feature
/kind flake
/kind other

What this PR does / why we need it:

How to verify it

Which issue(s) this PR fixes:

Special notes for your reviewer:

Does this PR introduce a user-facing change?

None

With all the recent speed-ups here the timeout is way to high, all tasks
should complete in under 30 mins generally. The smoke test in under
10min as it does not do much.

In particular I noticed at least two separate rootless integration tests
time out after 120min[1,2], obviously the tests do not take that long and
they are hanging somehwere instead. With a lower timeout we do not waste
so much time when this happens.

[1] https://cirrus-ci.com/task/4733420225429504
[2] https://cirrus-ci.com/task/5597909967699968

Signed-off-by: Paul Holzinger <[email protected]>
@Luap99
Copy link
Member Author

Luap99 commented Feb 19, 2025

@flouthoc @nalind PTAL

Copy link
Collaborator

@flouthoc flouthoc left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM
/lgtm

@flouthoc
Copy link
Collaborator

@Luap99 Also I am curious any hints why this hangs are happening ? I have seen them a few times when certain tests are just hanged.

@flouthoc
Copy link
Collaborator

Often in these hangs the logs are also not printend.

@nalind
Copy link
Member

nalind commented Feb 19, 2025

/approve

Copy link
Contributor

openshift-ci bot commented Feb 19, 2025

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: flouthoc, Luap99, nalind

The full list of commands accepted by this bot can be found here.

The pull request process is described here

Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@openshift-merge-bot openshift-merge-bot bot merged commit 318f35e into containers:main Feb 19, 2025
34 checks passed
@Luap99 Luap99 deleted the cirrus-timeout branch February 19, 2025 18:24
@Luap99
Copy link
Member Author

Luap99 commented Feb 19, 2025

@flouthoc yeah well that is from someone to figure out why things hang, of course if a bats test hangs no further logs will be printed. It is also possible but some of the recent log lines are still buffered so it is not like the log exactly tells you which line hangs.

Best bet try to reproduce outside of CI or log at the logs in CI, we see to capture the journald output so maybe you find clues there. Let me file an issue and we can debug there.

@Luap99
Copy link
Member Author

Luap99 commented Feb 19, 2025

#5996

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants