Hi Community!
I have a problem with a big Logfile. This log
If i make a realtimesearch for that specific source, some events are disappeared.
I have recorded some of this missing events and found it later in the index with a delay of more than 2 hours.
At my indexer I see sometimes the following error for that sourcetype
AggregatorMiningProcessor - Too many events (300K)
It looks like the universalforwarder doesn't sent new events to the indexer and after a while a hugh load would be send.
Do you have an idea what I can do now?
Thanks
Rob
So, you say you have ~250 events per minute (that's nothing special by the way, I've seen much more talkative log files), but splunk is complaining about 300 thousand events with the same timestamp? Sounds like something is seriously broken with timestamping somehow?
You mention you struggle parsing the date due to the German month names. Can you try just parsing the time? If I'm not mistaken, Splunk will default to the current date, if you only extract the time.
Can you post full error event from _internal logs. Also this may help: http://docs.splunk.com/Documentation/Splunk/7.1.0/Data/Resolvedataqualityissues
I have seen that post, but I'm using 6.4.4 and didn't find that data quality dashboard in the dmc
Is this the complete error message?
05-15-2018 12:57:49.175 +0200 WARN AggregatorMiningProcessor - Too many events (300K) with the same timestamp: incrementing timestamps 3 second(s) into the future to insure retrievability - data_source="/opt/myapp/myapp.log", data_host="machine1", data_sourcetype="myapp"
I haven't found a solution for transforming my german timestamp
/* Di Mai 15 2018 10:42:02.9290 */
into the event timestamp but normaly , if all events were sent to the indexer, the events are near realtime, but this is an other thing.
As @FrankVI also said, your problem is most likely based on broken timestamp recognition.
Do you read that file locally, or is it send to you via syslog?