<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Dataset best practices in Splunk Enterprise</title>
    <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746191#M22252</link>
    <description>&lt;P&gt;Hello!&lt;/P&gt;&lt;P&gt;I maintain Splunk reports. Some of the Pivot reports are based on a Dataset that is generated based on a simple search. Duplicate values ​​have not been taken into account in the generation.&lt;/P&gt;&lt;P&gt;Due to an error, there were two data sources for a few weeks. This resulted in identical duplicate rows in the dataset.&lt;/P&gt;&lt;P&gt;In the future, duplicate rows can be removed from the dataset with a simple dedup. However, are there any best practices to fix this?&lt;/P&gt;</description>
    <pubDate>Wed, 14 May 2025 14:15:39 GMT</pubDate>
    <dc:creator>RdomSplunkUser7</dc:creator>
    <dc:date>2025-05-14T14:15:39Z</dc:date>
    <item>
      <title>Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746191#M22252</link>
      <description>&lt;P&gt;Hello!&lt;/P&gt;&lt;P&gt;I maintain Splunk reports. Some of the Pivot reports are based on a Dataset that is generated based on a simple search. Duplicate values ​​have not been taken into account in the generation.&lt;/P&gt;&lt;P&gt;Due to an error, there were two data sources for a few weeks. This resulted in identical duplicate rows in the dataset.&lt;/P&gt;&lt;P&gt;In the future, duplicate rows can be removed from the dataset with a simple dedup. However, are there any best practices to fix this?&lt;/P&gt;</description>
      <pubDate>Wed, 14 May 2025 14:15:39 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746191#M22252</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-14T14:15:39Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746199#M22253</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.splunk.com/t5/user/viewprofilepage/user-id/310168"&gt;@RdomSplunkUser7&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I think ultimately this depends on what your searches are doing, if there is a risk of pulling in duplicate data then dedup is a good option, or you could look at using something like &lt;EM&gt;stats latest(fieldName) as latestFieldName&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;It really depends on your search(es). If you'd like to share the SPL we might be able to help further.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-unicode-emoji" title=":glowing_star:"&gt;🌟&lt;/span&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;Did this answer help you?&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;If so, please consider:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Adding karma to show it was useful&lt;/LI&gt;&lt;LI&gt;Marking it as the solution if it resolved your issue&lt;/LI&gt;&lt;LI&gt;Commenting if you need any clarification&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Your feedback encourages the volunteers in this community to continue contributing&lt;/P&gt;</description>
      <pubDate>Wed, 14 May 2025 16:11:11 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746199#M22253</guid>
      <dc:creator>livehybrid</dc:creator>
      <dc:date>2025-05-14T16:11:11Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746258#M22260</link>
      <description>&lt;P&gt;It is based on very simple search.&lt;/P&gt;&lt;P&gt;index=&amp;lt;index_name&amp;gt; sourcetype= &amp;lt;blaahaa&amp;gt;&amp;nbsp; field2. After this, a number of fields are extracted using rex.&lt;/P&gt;&lt;P&gt;I would like to include in the search as a new contrain a&amp;nbsp; very simple dedup clause "| dedup _raw|".&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is this advisable?&lt;/P&gt;</description>
      <pubDate>Thu, 15 May 2025 14:45:47 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746258#M22260</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-15T14:45:47Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746317#M22266</link>
      <description>&lt;P&gt;I have an identical situation as described here &amp;lt;&lt;A href="https://community.splunk.com/t5/Reporting/How-to-not-include-the-duplicated-events-while-accelerating-the/m-p/244884" target="_blank"&gt;https://community.splunk.com/t5/Reporting/How-to-not-include-the-duplicated-events-while-accelerating-the/m-p/244884&lt;/A&gt;&amp;gt;&lt;/P&gt;</description>
      <pubDate>Fri, 16 May 2025 12:37:27 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746317#M22266</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-16T12:37:27Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746318#M22267</link>
      <description>&lt;P&gt;Here are some guidance how to resolve the problem&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.splunk.com/t5/Splunk-Search/How-to-delete-duplicate-events/td-p/70656?_gl=1*va2tii*_gcl_au*MTgyNjg5MzM4NS4xNzQ3MDQ5MDAyLjE4MjI0OTQwOTkuMTc0NzIzMTA3Ni4xNzQ3MjMxMDc1*FPAU*MTgyNjg5MzM4NS4xNzQ3MDQ5MDAy*_ga*MTE2NjU0NjgxNC4xNzQ3MDQ5MDAy*_ga_5EPM2P39FV*czE3NDczOTczMzQkbzkkZzEkdDE3NDczOTkwNTckajAkbDAkaDIwMzUxMzI1NzU.*_fplc*ZWZ1MWJ5V3h4NFVZd1ZpMVJqc2xKOU1WdHo2WTNQdU1OcUlhWlE3bGpXdXU3ZENuYjJFOXppSDNCSVRvcENOcUxuaUpWSU5FUkpGaXFNMG9DN0slMkYlMkZKdUtPVWZncHhEY1lmUDQlMkI1RFJGU2NOVmhYaDFJSlpxMWszNDRHbDB3JTNEJTNE" target="_blank"&gt;https://community.splunk.com/t5/Splunk-Search/How-to-delete-duplicate-events/td-p/70656?_gl=1*va2tii*_gcl_au*MTgyNjg5MzM4NS4xNzQ3MDQ5MDAyLjE4MjI0OTQwOTkuMTc0NzIzMTA3Ni4xNzQ3MjMxMDc1*FPAU*MTgyNjg5MzM4NS4xNzQ3MDQ5MDAy*_ga*MTE2NjU0NjgxNC4xNzQ3MDQ5MDAy*_ga_5EPM2P39FV*czE3NDczOTczMzQkbzkkZzEkdDE3NDczOTkwNTckajAkbDAkaDIwMzUxMzI1NzU.*_fplc*ZWZ1MWJ5V3h4NFVZd1ZpMVJqc2xKOU1WdHo2WTNQdU1OcUlhWlE3bGpXdXU3ZENuYjJFOXppSDNCSVRvcENOcUxuaUpWSU5FUkpGaXFNMG9DN0slMkYlMkZKdUtPVWZncHhEY1lmUDQlMkI1RFJGU2NOVmhYaDFJSlpxMWszNDRHbDB3JTNEJTNE&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 16 May 2025 12:41:11 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746318#M22267</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-16T12:41:11Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746326#M22268</link>
      <description>IMHO: I don't like or suggest you to add delete permissions to anyone permanently! It isn't great idea to run scheduled job which are removing events from splunk.</description>
      <pubDate>Fri, 16 May 2025 13:54:41 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746326#M22268</guid>
      <dc:creator>isoutamo</dc:creator>
      <dc:date>2025-05-16T13:54:41Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746327#M22269</link>
      <description>&lt;P&gt;I dont like the idea you can not add dedup with pipelines in the simple base search of dataset.&amp;nbsp;&lt;/P&gt;&lt;P&gt;The Splunk should offer a ready method to deduplicate index.&lt;/P&gt;</description>
      <pubDate>Fri, 16 May 2025 13:58:53 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746327#M22269</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-16T13:58:53Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746329#M22270</link>
      <description>&lt;P&gt;Unfortunately at least I didn't know any generic answer for this.&lt;/P&gt;&lt;P&gt;That method what they presented here is one option, but as said you need to be 100% sure that it works with your data and test it several times to be sure!&lt;/P&gt;&lt;P&gt;And of course you must 1st get rid of those new duplicates and ensure that all your inputs works as they should without duplicating new events.&lt;/P&gt;&lt;P&gt;After that you probably could do that delete if you are absolutely sure that it works also in your case. And I propose that you should use some temp account which has can_delete role just for this time what is needed to do that clean up.&lt;/P&gt;</description>
      <pubDate>Fri, 16 May 2025 14:10:07 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746329#M22270</guid>
      <dc:creator>isoutamo</dc:creator>
      <dc:date>2025-05-16T14:10:07Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746330#M22271</link>
      <description>Of course you can add dedup on those queries but this will kill your performance!&lt;BR /&gt;And it depends how this duplication has happened and how you could identified those events? That just depends on how that has happened, can you e.g. dedup _raw or just set of different fields or did it needs some calculations/modifications (e.g. times) too?</description>
      <pubDate>Fri, 16 May 2025 14:12:43 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746330#M22271</guid>
      <dc:creator>isoutamo</dc:creator>
      <dc:date>2025-05-16T14:12:43Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746364#M22279</link>
      <description>&lt;P&gt;Maybe I can make a new data set which is the original data set minus duplicate identical log lines. Are there any tutorials for this? I am newbie super user for just for some reports. I hate this role.&lt;/P&gt;</description>
      <pubDate>Fri, 16 May 2025 19:37:38 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746364#M22279</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-16T19:37:38Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746366#M22280</link>
      <description>How big and complex your dataset is and how much its content is changing? And how long time span it covers?</description>
      <pubDate>Fri, 16 May 2025 19:46:53 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746366#M22280</guid>
      <dc:creator>isoutamo</dc:creator>
      <dc:date>2025-05-16T19:46:53Z</dc:date>
    </item>
    <item>
      <title>Re: Dataset best practices</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746371#M22281</link>
      <description>&lt;P&gt;My situation is quite simple. I lack the basic training. The ChatGPT showed me the way. Maybe this can be an answer for another Splunk report newbie super user.&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;”Ah, so you want to deduplicate data and use it in a Pivot table – great clarification! Pivot in Splunk is based on Data Models, and indeed, there are limitations on SPL commands (like | dedup _raw) in that context.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":magnifying_glass_tilted_left:"&gt;🔍&lt;/span&gt; The Problem:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Pivot uses a Data Model, and in the base search of a Data Model, you cannot use pipe (|) commands like dedup.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":direct_hit:"&gt;🎯&lt;/span&gt; Your Goal:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Remove duplicates based on _raw and still use the data in a Pivot table.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":white_heavy_check_mark:"&gt;✅&lt;/span&gt; Solution Options for Use in Pivot:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":repeat_button:"&gt;🔁&lt;/span&gt;&lt;SPAN class=""&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;1. Create a Saved Search with dedup, then build a Data Model on top of it&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;This is the recommended method:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Step 1: Create a Saved Search&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;SPAN&gt;Go to Splunk’s Search view.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Enter the SPL:&lt;/SPAN&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;index=your_index sourcetype=your_sourcetype&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;| dedup _raw&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Click Save As → Report.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Name it, for example: Deduped Raw Events.&lt;/SPAN&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Step 2: Create a new Data Model based on that Saved Search&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;SPAN&gt;Go to Settings → Data Models → New Data Model.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Give it a name and save it.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Add a new Object, and choose:&lt;/SPAN&gt;&lt;SPAN&gt;&lt;BR /&gt;&lt;/SPAN&gt;&lt;/LI&gt;&lt;UL&gt;&lt;LI&gt;&lt;SPAN&gt;Object Type: Event&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Constraint: Use the saved search:&lt;/SPAN&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;LI&gt;&amp;nbsp;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;savedsearch="Deduped Raw Events"&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&amp;nbsp;&lt;/LI&gt;&lt;UL&gt;&lt;LI&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;NOTE: savedsearch="your_report_name" references the saved search.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&amp;nbsp;&lt;/LI&gt;&lt;/UL&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Step 3: Use Pivot on top of this Data Model&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;SPAN&gt;Go to Pivot → Select your new Data Model → Deduped Raw Events.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Build your table as desired.&lt;/SPAN&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":warning:"&gt;⚠️&lt;/span&gt; Notes:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;SPAN&gt;This only works if the saved search is public (shared) or you have permission to use it.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;The Saved Search must return fields that you can use in Pivot (like _time, host, source, custom fields, etc.).&lt;/SPAN&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;🧪 Option 2: Simulate Dedup within the Data Model (if possible)&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Data Models do not allow | dedup, but you can:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;SPAN&gt;Add an auto-extracted field, which lets you group by that field in Pivot.&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;Or, if you have a unique identifier (e.g., event_id), you can use first-value or latest-value aggregations in Pivot to simulate deduplication.&lt;/SPAN&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":pushpin:"&gt;📌&lt;/span&gt; Summary:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;TABLE cellspacing="0" cellpadding="0"&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;Method&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;Dedup Allowed?&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;Usable in Pivot?&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;Saved Search + dedup&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":white_heavy_check_mark:"&gt;✅&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":white_heavy_check_mark:"&gt;✅&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;Native Data Model search&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":cross_mark:"&gt;❌&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":white_heavy_check_mark:"&gt;✅&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;SPL with pipes in Pivot UI&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":cross_mark:"&gt;❌&lt;/span&gt; (not allowed)&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD&gt;&lt;P&gt;&lt;SPAN&gt;&lt;span class="lia-unicode-emoji" title=":white_heavy_check_mark:"&gt;✅&lt;/span&gt; but very limited&lt;/SPAN&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;If you’d like, I can also help you write the full search or configure it for a specific type of data or log source – just let me know what you’re using it for in Pivot!”&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;The data is very simple event log type data. The amount of data is small. There is a unique field in log lines (event id). The question was about how to tweak existing data set. Splunk is not good for these type of business reports which should be moved to another report platform (ie MSBI).&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 16 May 2025 20:19:20 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise/Dataset-best-practices/m-p/746371#M22281</guid>
      <dc:creator>RdomSplunkUser7</dc:creator>
      <dc:date>2025-05-16T20:19:20Z</dc:date>
    </item>
  </channel>
</rss>

