-
Notifications
You must be signed in to change notification settings - Fork 678
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is there a health check of any kind on logspout?? #207
Comments
Is there no response from |
There was a response from docker logs in the cases that I saw. |
Also now watching this issue. I'm seeing similar behavior. Available and happy to help debug if you can suggest a place for me to start. Noticed I'm using an old version -- will be updating to latest release tomorrow and seeing what happens. |
I've been having (probably) the same problem for about 2 weeks. I've been collecting data otherwise I would have submitted the issue sooner. I'm using I will try and leave the broken pod in the current broken state so I can collect more information from it or the node, please send me any requests for info you have. When a container breaks, the most noticeable issue is that logs cease to be sent to papertrail. The container stays running. If I port forward to the pod so I can connect to the container's http server my kubectl port-forward logspout-papertrail-rsjok 9002:80
curl localhost:9002/logs # curl stays running, no output
The above commands work fine and output logs when I connect to a working logspout container. I can log into the kubernetes node and strace the logspout pid:
logspout just seems to be polling and not noticing any log events. |
The docker image ID of my current broken container is |
Thats interesting. Do you notice any other issues around the time it stops shipping data? We could/should probably add statsd support so we can pinpoint how much data/when things stop being shipped. I'd happily look at such a pr and merge it for our next release. |
@josegonzalez No other issues around the time. This cluster is a low-traffic cluster so it should be pretty easy to tell if something is spiking or otherwise not working correctly. |
In other words, I think this is an issue of basic functionality breaking, not a resource related edge case. |
Are you using latest master or something else? We may have fixed your problem here #191. |
I've been using |
@josegonzalez So I've switched to docker master tag (image id: The behavior I'm seeing is a small number of containers' logs get through. As I'm writing this only one container's logs, from one host, are getting into papertrail. If I connect directly via http to the logspout container I see the same behavior. If I use I have recently returned to papertrail at my new job and I'm surprised to be having these issues. I never had a single problem with the now deprecated |
I switched back to |
I am still seeing it intermittently unfortunately. Today I saw a new error though, I tried to curl to follow the log output of a container and got this response:
which was new. Restarting the logspout container fixed it though. We are considering pursuing an alternate method of using logback to write logs directly to kafka due to these issues unfortunately :/ |
@chrishiestand interesting. gliderlabs/logspout is essentially progrium/logspout, except further along in development. I'm not personally sure when things "broke" in such a way that we'd no longer be handling those cases, but unfortunately test coverage isn't great. There are a few PRs that should help ameliorate the issue, but they are still in a state of flux. If someone would like to fund the work to investigate/fix any such issues, I'd be happy to talk, but otherwise this will have to take a backseat to other, paid work that the maintainers have. |
From struggling with suddenly missing logs for the past several weeks, I'm quite sure at least one of the causes is log rotation when using the @edude03 give the |
I'm hitting this issue when using logspout on kubernetes. I switched to UDP instead of TCP and the issue seems to have been mitigated some. It's gone from seeming to stop within 24 hours to having been up for a few days. I'm wondering if it's volume related since I don't normally hit this on quieter machines, but the kubernetes setup I'm using is much more active. I had been using the |
Seems like this issue has meandered a bit and looks like a few things are discussed here. I'm going to close this and those that are still having a specific issue, please open a dedicated issue for that. Note: there is a known issue for log flows halting if using docker's |
Related to #216. |
I have created this gist to show how I solved this by checking if logspout is sending outgoing traffic (TCP or UDP, which are my current needs).. im using this as a swarm healthcheck |
I have logspout set to publish its logs to a port, and typically I can view its output by curling that port, however I am currently getting no response. If I restart the container, it will typically solve this problem, but I'm trying to find a root cause. Is there a way to health check an existing logspout container to see why it is not responding before I restart it??
The text was updated successfully, but these errors were encountered: