You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Mar 18 14:54:14 kubelet[35502]: E0318 14:54:14.509505 35502 container_log_manager.go:253] "Container log doesn't exist, reopen container log failed" err="rpc error: code = Unknown desc = docker does not support reopening container log files"
Mar 18 14:54:14 kubelet[35502]: E0318 14:54:14.512587 35502 remote_runtime.go:751] "ReopenContainerLog from runtime service failed" err="rpc error: code = Unknown desc = docker does not support reopening container log files"
The SPAM is real.... for a 10 node cluster we had over 200,000 log lines of this nature in 1 hour.
There are no "workarounds" for users who do not use the 'json-file' logging driver.
Suggested workaround, from #35
fixed when setting container-log-max-size to the same size as docker daemon max-size log opts.
Are users just subject to the pain?
I saw in one thread that using a stub logging manager with cri-dockerd could work. Is there any documentation to help implement this?
Not sure what version of docker this was added in, but current versions of docker default to dual-logging when using a remote logging driver (to facilitate the docker logs command, and likely kubectl logs). That is, they log to remote and local. This is pretty obscure in the docs and easy to miss as it's not well referenced from the remote logging driver pages. https://docs.docker.com/engine/logging/dual-logging/
Dual-logging can be disabled with the cache-disabled logging opt. Or, you can set cache-max-size (comparable to max-size when using local logging drivers, as referenced in #35), which will allow docker logs to still work.
One additional hint for anyone running across this: if setting the above in docker's daemon.json, you'll have to recreate all containers. The logging config for existing containers is not updated.
Expected Behavior
The SPAM is real.... for a 10 node cluster we had over 200,000 log lines of this nature in 1 hour.
There are no "workarounds" for users who do not use the 'json-file' logging driver.
Suggested workaround, from #35
Are users just subject to the pain?
I saw in one thread that using a stub logging manager with cri-dockerd could work. Is there any documentation to help implement this?
This is our daemon configuration.
Specifications
The text was updated successfully, but these errors were encountered: