Deployment Architecture

size of batches

MarkRichardsNH
New Member

Once a day we drop 20k files into a batch directory for processing.

It can take 2-3 hours for all of the files to be processed (by heavy forwarder) .

During that time the file count remains consistent and then when it is done, we see Splunk will purge the 20k files.

Is there something we can do to chunk this up into smaller batch sizes, so that for instance, it digests 1000 files, deletes them and then moves onto the next 1000 files . (or indeed deletes each processed file before digesting another one)

I ask because during this 2-3 hour window we need to make sure splunkd is not cycled, otherwise it will start again and create multiple duplicate transactions.

An alternative it to make these Monitor instead of batch, but these files won't change until 24 hours later and having 40k files being monitored (we have another set of files in a similar situation) seems to be a unnecessary burden on the system.

Tags (1)
0 Karma

DalJeanis
Legend

Interesting question, and interesting behavior by splunk. I don't see any specification in the docs regarding timing of deletion of batch files, so the alternatives I see are both extra-splunk... 1) multiple batch directories 2) load the batch directory in batches.

0 Karma

MarkRichardsNH
New Member

Thanks, I couldn't find anything obvious either, so if we have no better suggestions I'm going to have to look into one of those alternatives.

0 Karma
Get Updates on the Splunk Community!

Stay Connected: Your Guide to November Tech Talks, Office Hours, and Webinars!

🍂 Fall into November with a fresh lineup of Community Office Hours, Tech Talks, and Webinars we’ve ...

Transform your security operations with Splunk Enterprise Security

Hi Splunk Community, Splunk Platform has set a great foundation for your security operations. With the ...

Splunk Admins and App Developers | Earn a $35 gift card!

Splunk, in collaboration with ESG (Enterprise Strategy Group) by TechTarget, is excited to announce a ...