-
Notifications
You must be signed in to change notification settings - Fork 8.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multi instance Task Manager issues after 8.15 #197145
Comments
Pinging @elastic/response-ops (Team:ResponseOps) |
Regarding the I don't believe the fix mentioned above will solve the |
Thanks @mikecote for your reply. So if I understand this correctly it's a mostly cosmetic issue that'll be fixed soon, but the task manager is actually working fine in the mean time and I can just ignore it if I don't need a proper capacity estimation / health status reporting. If the expired warnings are unrelated and therefore something I gotta investigate further myself, this issue can be closed again then. Thanks a lot for the help :) |
That is correct, the calculations are based on the wrong number of observed Kibana instances so it's providing false warnings.
That is my thinking, look for "failed to poll for work" logs or others coming from the task manager plugin, it should help find the underlying cause. I'll leave the issue opened a bit longer just in case they end up being related in your case. |
I'll go ahead and close the issue now, I hope it the above helped! |
Kibana version:
8.15+
Elasticsearch version:
8.15+
Server OS version:
Docker on Debian
Browser version:
N/A
Browser OS version:
N/A
Original install method (e.g. download page, yum, from source, etc.):
Docker
Describe the bug:
I have seen multiple clusters that are throwing Taskmanager errors (Degraded, even though not overloaded at all, and HealthStatus.Error because of expired hot timestamps). Furthermore, they only show a single 'observed_kibana_instances' on api/task_manager/_health API Endpoints. However, on Stack Monitoring all kibana instances are shown.
I guess something is regularly killing my task managers on multiple instances, and somehow they don't appear to 'talk' to each other.
I haven't observed this before 8.15, and a cluster on 8.14 is still working fine (with pretty much identical config)
Steps to reproduce:
Expected behavior:
Multiple kibana instances should be shown on the health api, and task manager should not be regularly degraded (once every 1-2 minutesor so)
Screenshots (if relevant):
Errors in browser console (if relevant):
Provide logs and/or server output (if relevant):
Any additional context:
The text was updated successfully, but these errors were encountered: