<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Re-index a file and prevent duplicates in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139801#M28728</link>
    <description>&lt;P&gt;You might want to make a report of the record IDs you have in Splunk, then cull those from your input file. Then use &lt;A href="http://docs.splunk.com/Documentation/Splunk/6.0/Data/MonitorfilesanddirectoriesusingtheCLI"&gt;splunk add oneshot&lt;/A&gt; to import the file (or some other method).&lt;/P&gt;</description>
    <pubDate>Wed, 13 Nov 2013 01:31:21 GMT</pubDate>
    <dc:creator>jtrucks</dc:creator>
    <dc:date>2013-11-13T01:31:21Z</dc:date>
    <item>
      <title>Re-index a file and prevent duplicates</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139800#M28727</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;

&lt;P&gt;I have some CSV files which were indexed, but a proportion of the events were corrupted in the index. Each file has up to 1 million records. Is there a way to ask Splunk to re-index this file and to only index events that it doesn't current have? Each event has a unique record ID field.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Nov 2013 01:09:51 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139800#M28727</guid>
      <dc:creator>JeremyHagan</dc:creator>
      <dc:date>2013-11-13T01:09:51Z</dc:date>
    </item>
    <item>
      <title>Re: Re-index a file and prevent duplicates</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139801#M28728</link>
      <description>&lt;P&gt;You might want to make a report of the record IDs you have in Splunk, then cull those from your input file. Then use &lt;A href="http://docs.splunk.com/Documentation/Splunk/6.0/Data/MonitorfilesanddirectoriesusingtheCLI"&gt;splunk add oneshot&lt;/A&gt; to import the file (or some other method).&lt;/P&gt;</description>
      <pubDate>Wed, 13 Nov 2013 01:31:21 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139801#M28728</guid>
      <dc:creator>jtrucks</dc:creator>
      <dc:date>2013-11-13T01:31:21Z</dc:date>
    </item>
    <item>
      <title>Re: Re-index a file and prevent duplicates</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139802#M28729</link>
      <description>&lt;P&gt;I was kind of hoping for something a little less manual....&lt;/P&gt;</description>
      <pubDate>Wed, 13 Nov 2013 01:36:27 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139802#M28729</guid>
      <dc:creator>JeremyHagan</dc:creator>
      <dc:date>2013-11-13T01:36:27Z</dc:date>
    </item>
    <item>
      <title>Re: Re-index a file and prevent duplicates</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139803#M28730</link>
      <description>&lt;P&gt;An easier way might be to delete the events you have in your index now, clean the fishbucket, and just let Splunk reindex them.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Nov 2013 01:53:30 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139803#M28730</guid>
      <dc:creator>ShaneNewman</dc:creator>
      <dc:date>2013-11-13T01:53:30Z</dc:date>
    </item>
    <item>
      <title>Re: Re-index a file and prevent duplicates</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139804#M28731</link>
      <description>&lt;P&gt;Clean the fishbucket?&lt;/P&gt;</description>
      <pubDate>Wed, 13 Nov 2013 01:59:19 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139804#M28731</guid>
      <dc:creator>JeremyHagan</dc:creator>
      <dc:date>2013-11-13T01:59:19Z</dc:date>
    </item>
    <item>
      <title>Re: Re-index a file and prevent duplicates</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139805#M28732</link>
      <description>&lt;P&gt;Yes. This is what I do.&lt;/P&gt;

&lt;P&gt;Run the search that has the events you need to delete, I assume you don't want to delete the entire index. If you do, run the below command with the index name that you wish to wipe out, then clean _thefishbucket. Otherwise, run your search to find your events, then pipe that "| delete".&lt;/P&gt;

&lt;P&gt;cd out to the Splunk\bin directory. Type splunk stop. Then type splunk clean eventdata -index _thefishbucket&lt;/P&gt;

&lt;P&gt;Then type splunk start. The rest is automatic, assuming you have fixed the files.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Nov 2013 02:14:01 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Re-index-a-file-and-prevent-duplicates/m-p/139805#M28732</guid>
      <dc:creator>ShaneNewman</dc:creator>
      <dc:date>2013-11-13T02:14:01Z</dc:date>
    </item>
  </channel>
</rss>

