Ever since upgrading to Splunk5 I've had an issue where after a random amount of time Splunk will quit logging events from some of my hosts. I have 15+ servers sending normal syslogs and running a custom app that sends it's events on 515. When this issue happens on these couple hosts, Splunk continues to receive syslog messages, but quits seeing any events from our app.
I've used tcpdump on hosts and the Splunk server; events are still being received by the server.
The port is still up and listening, because I'm still getting the app data from the other hosts.
I've looked in splunkd.log, don't see anything happening around the time I quit getting events.
It's like Splunk has decided to just ignore these particular hosts, but only on the 515 port.
Restarting splunk fixes it and everything works, for about 10-12 hours..
I did have some SEDCMD commands, but I have since commented those out and I am still see the issue.
I'm not sure where else to look....
I would recommend contacting support, perhaps the fix does not quite cover all possible cases. Also it is a best practise to send your data into an rsyslog/syslog-ng first and then have splunk monitor those files.