-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Failing Test]: PythonPostCommit is Extremely Flaky #29214
Comments
Is this still relevant? |
Yep still a problem, @damccorm said there had been a few issues along these lines with the self-hosted runners |
Im looking into this one, there are issues with runners that soon will be fixed but that is still not the final fix needed for python postcommit, keep you posted |
.take-issue |
@volatilemolotov I don't think you meant to auto-close this with the PR, is that right? If yes, we can reclose after a green signal anyways I guess |
I did not meant to autoclose its only a part. Sorry im not aware of mechanisms, worked on a lot of different systems :) |
No worries, its actually a GitHub feature - https://docs.github.com/en/issues/tracking-your-work-with-issues/linking-a-pull-request-to-an-issue#linking-a-pull-request-to-an-issue-using-a-keyword |
So a scheduled run failed with one actually passing. Other three jobs fail in different places Any ideas what is going on? Could it be because of parallel run (I had a full green run in my fork) |
Definitely seems like we've upgraded from permared jobs to test flakiness, so I don't think this is a runner/actions problem anymore. For example https://github.com/apache/beam/actions/runs/6797617071/job/18480096798 already has 3 green jobs (with a 4th still running) At least some of it is caused by #29076 - I see a bunch of failures related to that test in the workflow you linked. I have #29197 to fix that, was holding off on merging since there was a lot going on causing issues, but it might be time to merge. I'm running https://github.com/apache/beam/actions/runs/6802293125 to make sure I'm correctly sickbaying it, but once that runs (assuming its working as expected) I think we should merge the PR. |
The flow you referenced is green So yeah, flakiness. Glad to have it sorted we were lucky that MTU issues did not cause bigger problems |
Now it is still flaky though with lower frequency: https://github.com/apache/beam/runs/20274778848 There is other flaky test. e.g.
|
There's one week until the 2.54.0 cut and this issue is tagged for that release, if possible/necessary please complete the necessary work before then, or move this to the 2.55.0 Release Milestone. This one seems like we may need to cherry pick though if additional fixes occur. |
Its still flaky, failing differently https://github.com/apache/beam/actions/runs/7563095321/job/20594841089#step:9:37337 Cloud be the test is somehow broken but i cannot see a pattern right now |
Closing this one, as all the python postcommits are passing on the release branch. |
What happened?
The apache_beam/io/external/xlang_kinesisio_it_test.py::CrossLanguageKinesisIOTest::test_kinesis_write test is failing in the Python PostCommit with a consistent error message:
The test is defined here: https://github.com/apache/beam/blob/master/sdks/python/apache_beam/io/external/xlang_kinesisio_it_test.py#L94
Specifically, the failure is in
create_stream()
:Issue Failure
Failure: Test is continually failing
Issue Priority
Priority: 1 (unhealthy code / failing or flaky postcommit so we cannot be sure the product is healthy)
Issue Components
The text was updated successfully, but these errors were encountered: