Deployment Architecture

Hadoop Data Roll and archiving replicated buckets

delappml_2
Explorer

I have an indexer cluster with a replication factor of 3. If I were to implement Hadoop Data Roll, would only one copy of each event be archived to Hadoop at freeze time, or would all three bucket copies be archived? I'm trying to find out if I can save in terms of raw archive storage costs by implementing HDR versus archiving frozen buckets to a set of NFS mounts.

1 Solution

rdagan_splunk
Splunk Employee
Splunk Employee

HDR will only copy 1 journal.gz (raw data from the bucket). Therefore, 3X Splunk bucket replication will not impact the storage on the Hadoop side.

View solution in original post

rdagan_splunk
Splunk Employee
Splunk Employee

HDR will only copy 1 journal.gz (raw data from the bucket). Therefore, 3X Splunk bucket replication will not impact the storage on the Hadoop side.

mattymo
Splunk Employee
Splunk Employee

HDR is by far the best archiving solution, unless you really want to write your own dedup logic (spoiler: you don't lol).

- MattyMo
Get Updates on the Splunk Community!

What's new in Splunk Cloud Platform 9.1.2312?

Hi Splunky people! We are excited to share the newest updates in Splunk Cloud Platform 9.1.2312! Analysts can ...

What’s New in Splunk Security Essentials 3.8.0?

Splunk Security Essentials (SSE) is an app that can amplify the power of your existing Splunk Cloud Platform, ...

Let’s Get You Certified – Vegas-Style at .conf24

Are you ready to level up your Splunk game? Then, let’s get you certified live at .conf24 – our annual user ...