We are a new SplunkCloud customer and are building out our instance, setting up our indexes, field extractions, etc. I’m currently working on Lookups and and seeing unexpected performance characteristics from the searches I am running.
I created an automatic lookup that links the data in one of our indexes to a lookup table that has about 15k rows and 7 columns of data. The automatic lookup links the index to the lookup table via a “guid" field.
This search takes 48 seconds to complete and has a scan count of 16million
This search takes 300ms to complete and has a scan count of 410
Why is the first search doing all of this extra work? We are about to roll out access to Splunk to about 150 employees. I want to make sure I understand the proper way to recommend people to run searches against this index that is linked to the lookup table.
The second search is going directly for the index to look for your guid. The first one needs to match (don't know how many fields are you matching against the lookup) every event to the lookup before filtering, and that's where this questions come into place
Is that lookup CSV or KVStore based?
If it is KVStore, maybe you'd prefer it to be replicated to the Indexer layer for performance increase?