-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How many clusters the daemon can manage #70
Comments
Strange, as I have used to monitor over 60 clusters with one daemon running. |
Thanks for your confirmation... its look like one cluster return an error and the daemon hang (config have to be checked) 2019-10-16 09:18:31,920:urllib3.connectionpool:WARNING: Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fa92 but wit the current configuration after removed the cluster on error we can see this on the log |
2019-10-16 09:28:57,264:urllib3.connectionpool:WARNING: Connection pool is full, discarding connection |
This is a concern with the way the collector schedules collection from multiple clusters (one problematic cluster can have knock-on effects). I have rewritten the collector in Go and plan to push the new version to github shortly. I intend to update the Python collector to Python 3 and to leverage the async functionality, but currently, the Go collector is more reliable and better-performing. |
Thks .. look forward to the new version in Go bye thks again |
I still haven't posted the Go collector externally (I will, I promise), but as noted above, the Python collector is certainly not limited to 8 clusters (we're using internally for close to 20). It is rather fragile wrt errors at startup however, and I need to fix that. In this case, the startup error is because it can't resolve the cluster name (but the urllib3 errors are terrible and don't actually tell you the name part of the URL). The connectionpool warnings are annoying (I opened a bug for those), but harmless in that they do not prevent the collector working (it just means that there are more connections than the cached limited so the collector is having to reconnect more frequently than we would like). |
Hello Tim |
FYI, the Golang collector is available at https://github.com/tenortim/gostats |
The daemon is not able to start when the number of clusters is superior at 8.
How can i manage that by using 2 deamons and 2 configuratin files?
The text was updated successfully, but these errors were encountered: