-
-
Notifications
You must be signed in to change notification settings - Fork 748
WIP: investigating deadlock #4435
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Also, check that we don't release keys with dependents
|
Tests seem to pass. I'm going to rerun them a couple of times. |
|
cc @fjetter |
|
Even the failures on travis look promising. I'll give it a spin on some real world stress situations and report back |
|
There are still failures popping up connected to the dependency fetching (Worker becomes a zombie afterwards) Haven't investigated, yet, but I suspect this key (or a related one) was stolen and the I tried at some point to modify distributed/distributed/worker.py Lines 2297 to 2298 in 467bbc2
to only delete the task from the internal dict if there are no more dependents, similar to the way Worker.data is handled a few lines above. This all led me to questioning the behaviour of Worker.release_key in the first place
|
|
Similarly, in a different run, I receive an error in Details |
|
I believe that @fjetter has continued this on his PR. Closing. |
This is a fork of #4432 . I'm just raising a PR in order to use CI for testing in a low-battery situation. Please ignore.