-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adopt common timeout/retry/backoff strategy for system tests. #1619
Comments
In python-docs-samples we're using a combination of flaky and not running flaky tests on travis (we have a jenkins instance). |
#1391 is related. |
I have an old branch around from #535 when I made the first stab at this. |
Maybe one of the following libraries would be useful: |
So now, in the systems-tests path we have from retry import Retry
@Retry(SomeException, tries=3, delay=30)
def method_to_retry():
... |
I added classes to handle two other cases in #2050:
|
@tseaver What about this issue? And the novel failures? |
I think we should create separate issues for the failures we still see, and resolve them one by one as we add backoff/retry for them. E.g., today's BQ failure should be its own issue, rather than reopening #1104. Likewise todays happybase failure should be a separate issue. |
@tseaver 500, 502 and 503 are just unavoidable for a general service. |
From #1609 (comment)
We have a number of sporadic system test failures (#1085, #1089, #1100, #1104) due to "eventual consistency" in the tested APIs. We have adopted a set of ad-hoc (and obviously not 100% reliable) strategies for working around the delays involved.
The point of this issue is to adopt a consistent policy / mechanism across all system tests for addressing delays caused by eventual consistency.
@jonparrott can you comment here, collecting your thoughts expressed elsewhere?
The text was updated successfully, but these errors were encountered: