Splunk Enterprise

Re-ingesting partial logs from one host to another

splunky1
Loves-to-Learn Everything

Our Splunk architecture is like

Two HFs pointing to Two internal Indexers and Two external Indexers.

Internal Indexers have different data and external indexers have different data (these indexers also receive data from other external HFs too) and HF's route the data correctly into the respective indexers.

We had a situation with fail over script which removed the outputs.conf file in both the HFs which resulted indexing the data locally into the HF and able to search this data in HFs but didn't go to indexers to search from search heads.

After we putting back the outputs.conf file the new data is going into the right indexers as intended but the data between is indexed into HF and lost in the indexers.

How can I re-ingest this data that's indexed in the HF into the indexers using the correct config.

I tried renaming the fish buckets folder and checked if that re-ingests the data but it only ingested small amount of data not everything.

I can still see data in my HF under $Splunk_home/var/lib/splunk/<index_name>/db/_raw 

What's the best way to re-ingest this data without manually moving the files into the indexers.

Thanks

0 Karma

aasabatini
Motivator

Hi @splunky1 

Good Challange!

anyway some considerations

  • if you have cluster indexer, move manually hte bucket is not great idea becuase you can create conflict with the bucket ID
  • The source is present on the HF or come from to  a UF?

Options:

  1.  Route data to splunk in syslog like is a third party system( cold be you will find to separate sourcetype)  https://docs.splunk.com/Documentation/Splunk/latest/Forwarding/Forwarddatatothird-partysystemsd
  2. If you have access on  your monitor stanza (inputs.conf) and  try force to reindex data with this setting
crcSalt = <SOURCE>

https://splunkonbigdata.com/2018/10/13/how-crc-works-in-splunk/

  • Try to summarize the logs with a schedule search (I know this is the more dirty option)
“The answer is out there, Neo, and it’s looking for you, and it will find you if you want it to.”
0 Karma
Get Updates on the Splunk Community!

SOC4Kafka - New Kafka Connector Powered by OpenTelemetry

The new SOC4Kafka connector, built on OpenTelemetry, enables the collection of Kafka messages and forwards ...

Your Voice Matters! Help Us Shape the New Splunk Lantern Experience

Splunk Lantern is a Splunk customer success center that provides advice from Splunk experts on valuable data ...

Building Momentum: Splunk Developer Program at .conf25

At Splunk, developers are at the heart of innovation. That’s why this year at .conf25, we officially launched ...