Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Controller worker thread times out during config set #66

Open
Cryptophobia opened this issue Mar 20, 2018 · 6 comments
Open

Controller worker thread times out during config set #66

Cryptophobia opened this issue Mar 20, 2018 · 6 comments
Labels

Comments

@Cryptophobia
Copy link
Member

From @kmala on April 21, 2016 18:10

when we try to do a config:set on an app with 30-40 pods the controller worker thread is timing out because the operation is taking more than 20min(default timeout of the worker thread) keeping the cluster in an unstable state with pods of both releases.

Copied from original issue: deis/controller#652

@Cryptophobia
Copy link
Member Author

From @bacongobbler on April 21, 2016 18:56

tagging at rc1 since this is something we need to fix before we cut a stable release.

@Cryptophobia
Copy link
Member Author

From @bacongobbler on May 19, 2016 18:30

ping @helgi; has this been fixed recently?

@Cryptophobia
Copy link
Member Author

From @helgi on May 19, 2016 18:32

No, and it won't be done in RC. This isn't a kubernetes problem, it's more the fact we are trying to execute any operation within the timeout of the gunicorn server, which we can't extend too much without causing contention of resources on the RC=1 controller setup. We'd need to move to background jobs to fix this

@Cryptophobia
Copy link
Member Author

From @helgi on May 19, 2016 18:33

Oh, what has helped is doing deploys in batches, and by default rolling as many pods as available nodes but that only mitigates the issue in some scenarios

@Cryptophobia
Copy link
Member Author

From @bacongobbler on May 19, 2016 19:20

@kmala are you still able to reproduce this in your environment with workflow-dev? If it's a matter of waiting for all the pods to come up then perhaps we should rethink this deployment strategy eventually, since the old way of just destroying everything and starting everything worked for 99% of our use cases, including this one. Graceful rolling deploys are nice but if it's causing us to be in an unstable state any time we're dealing with a larger number of jobs then perhaps we should go back to square one.

@Cryptophobia
Copy link
Member Author

From @mboersma on September 6, 2016 20:25

This situation should be improved by the batching operation of the current controller, although probably not fixed definitively.

duanhongyi added a commit to duanhongyi/controller that referenced this issue Apr 24, 2023
…ry-5.2.2

chore(deps): bump celery from 5.1.2 to 5.2.2 in /rootfs
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant