Knowledge Management

Spunk forwarder scalability

dimitarvalov
Engager

I'm considering usage of splunk-forwarder to integrate a system that generates many small files that contain log messages, i.e. at times more than a thousand per second. Once the files reach splunk they can be deleted.

I wonder how will the forwarded handle this situation. I've read that it can monitor well about a 100 files. Should I implement other jobs to move the processed files and how should I know if a file is processed?

The other approach that I could take is to change the system to log a rotating file. So which one do you think is better?

0 Karma
1 Solution

echalex
Builder

I would say prefer to log to a rotating file. I have experience from Splunking an application that was producing several thousand files an hour and it was not pretty. Monitoring that amount of files will result in a performance hit which is disproportionate to the amount of data to be ingested.

Therefore: prefer fewer log files.

View solution in original post

echalex
Builder

I would say prefer to log to a rotating file. I have experience from Splunking an application that was producing several thousand files an hour and it was not pretty. Monitoring that amount of files will result in a performance hit which is disproportionate to the amount of data to be ingested.

Therefore: prefer fewer log files.

Get Updates on the Splunk Community!

October Community Champions: A Shoutout to Our Contributors!

As October comes to a close, we want to take a moment to celebrate the people who make the Splunk Community ...

Community Content Calendar, November Edition

Welcome to the November edition of our Community Spotlight! Each month, we dive into the Splunk Community to ...

Stay Connected: Your Guide to November Tech Talks, Office Hours, and Webinars!

What are Community Office Hours? Community Office Hours is an interactive 60-minute Zoom series where ...