Splunk Search

What are some solutions for high cardinality field reporting?

the_wolverine
Champion

We have high cardinality data -- virtually every event is unique except for a small percentage of cases that we care about. So we're finding that we have count the unique ids, track them somehow in order to find the duplicates. Its just not feasible in Splunk when we have millions of events per minute.

Example:

search | stats count by unique_id | where count>1

(Millions of events per minute, results in a few hundred events where count>1). Summary indexing is not really a solution here since the unique_id could cross time/minute boundary.

0 Karma

valiquet
Contributor

|fields unique_id
|stats list( unique_id) AS id
| where mvfilter(mvcount(id)>1) != NULL

0 Karma

somesoni2
Revered Legend

May be setup indexed time field extraction for your unique_id field so you could use tstats with it.

Get Updates on the Splunk Community!

Accelerating Observability as Code with the Splunk AI Assistant

We’ve seen in previous posts what Observability as Code (OaC) is and how it’s now essential for managing ...

Integrating Splunk Search API and Quarto to Create Reproducible Investigation ...

 Splunk is More Than Just the Web Console For Digital Forensics and Incident Response (DFIR) practitioners, ...

Congratulations to the 2025-2026 SplunkTrust!

Hello, Splunk Community! We are beyond thrilled to announce our newest group of SplunkTrust members!  The ...