<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: How to avoid duplicate events when using distributed search across cloned indexers in Splunk Search</title>
    <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12097#M914</link>
    <description>&lt;P&gt;Thanks for the responses. The idea of including the name or IP of the indexer in the searches works as long as you are only doing ad-hoc searches and not if you have scheduled searches and reports. The LB option works but it increases cost and becomes a little more challenging if the indexers are geographically dispersed. &lt;/P&gt;

&lt;P&gt;Seems like this feature may be a candidate RFE. &lt;/P&gt;</description>
    <pubDate>Sun, 25 Apr 2010 18:06:41 GMT</pubDate>
    <dc:creator>Phil_T_</dc:creator>
    <dc:date>2010-04-25T18:06:41Z</dc:date>
    <item>
      <title>How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12094#M911</link>
      <description>&lt;P&gt;I have a scenario where A and B are indexers with one being the clone of the other. The idea being A is in one data center and B in a DR data center. &lt;/P&gt;

&lt;P&gt;C is a search head and if it is configured with both A and B as search peers, the searches return duplicates. &lt;/P&gt;

&lt;P&gt;Is there a way to configure C to pull data only from A and automatically start using B if A fails? &lt;/P&gt;

&lt;P&gt;Phil&lt;/P&gt;</description>
      <pubDate>Fri, 23 Apr 2010 01:30:52 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12094#M911</guid>
      <dc:creator>Phil_T_</dc:creator>
      <dc:date>2010-04-23T01:30:52Z</dc:date>
    </item>
    <item>
      <title>Re: How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12095#M912</link>
      <description>&lt;P&gt;There currently is not without using an external load balancer TCP to alias and failover A and B. Note you probably will have to modify A and B to have the same Splunk search node name (&lt;CODE&gt;serverName&lt;/CODE&gt; in &lt;CODE&gt;server.conf&lt;/CODE&gt;).&lt;/P&gt;</description>
      <pubDate>Fri, 23 Apr 2010 10:12:00 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12095#M912</guid>
      <dc:creator>gkanapathy</dc:creator>
      <dc:date>2010-04-23T10:12:00Z</dc:date>
    </item>
    <item>
      <title>Re: How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12096#M913</link>
      <description>&lt;P&gt;The LB play is doable but in lieu of that you can prepend all of your searches with splunk_server=A.  If it yells at you that A isn't available you can then search B.  &lt;/P&gt;</description>
      <pubDate>Fri, 23 Apr 2010 12:24:35 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12096#M913</guid>
      <dc:creator>dskillman</dc:creator>
      <dc:date>2010-04-23T12:24:35Z</dc:date>
    </item>
    <item>
      <title>Re: How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12097#M914</link>
      <description>&lt;P&gt;Thanks for the responses. The idea of including the name or IP of the indexer in the searches works as long as you are only doing ad-hoc searches and not if you have scheduled searches and reports. The LB option works but it increases cost and becomes a little more challenging if the indexers are geographically dispersed. &lt;/P&gt;

&lt;P&gt;Seems like this feature may be a candidate RFE. &lt;/P&gt;</description>
      <pubDate>Sun, 25 Apr 2010 18:06:41 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12097#M914</guid>
      <dc:creator>Phil_T_</dc:creator>
      <dc:date>2010-04-25T18:06:41Z</dc:date>
    </item>
    <item>
      <title>Re: How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12098#M915</link>
      <description>&lt;P&gt;I would like to see this feature as well, seems simple enough to implement - just the reverse of auto-lb from forwarders to the indexers.&lt;/P&gt;</description>
      <pubDate>Thu, 20 May 2010 22:26:46 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12098#M915</guid>
      <dc:creator>cudgel</dc:creator>
      <dc:date>2010-05-20T22:26:46Z</dc:date>
    </item>
    <item>
      <title>Re: How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12099#M916</link>
      <description>&lt;P&gt;If I understand splunk data cloning correctly, you might not want to just point to one indexer for results and then failover. Say A goes down and now your using B(your DR).. well when A comes back online it does not have the data that B has nor will it ever.. So you will be missing out on data if you only are querying A.. &lt;/P&gt;

&lt;P&gt;unless your doing some tricky failover the nfs file system or some file system replicaton for failover and then starting splunk up using that 'A' dataset.. which has a bunch of issues in itself..&lt;/P&gt;</description>
      <pubDate>Tue, 10 Aug 2010 04:39:48 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12099#M916</guid>
      <dc:creator>chicodeme</dc:creator>
      <dc:date>2010-08-10T04:39:48Z</dc:date>
    </item>
    <item>
      <title>Re: How to avoid duplicate events when using distributed search across cloned indexers</title>
      <link>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12100#M917</link>
      <description>&lt;P&gt;You don't need to manually do this. You can let splunk search both indexers and dedup the data. Assuming you don't have many duplicate events with the same content, time stamp, host and source.&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;`your_search` | dedup host, source, _time, _raw, splunk_server
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;This way if an event is on both servers, you only get one.&lt;BR /&gt;
If either server is down, you get events from the other.&lt;BR /&gt;
If data only got indexed on one it doesn't get missed.&lt;/P&gt;

&lt;P&gt;Bob&lt;/P&gt;</description>
      <pubDate>Wed, 27 Apr 2011 13:41:38 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/How-to-avoid-duplicate-events-when-using-distributed-search/m-p/12100#M917</guid>
      <dc:creator>BobM</dc:creator>
      <dc:date>2011-04-27T13:41:38Z</dc:date>
    </item>
  </channel>
</rss>

