<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: How archiving splunk logs to hdfs work in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172783#M34770</link>
    <description>&lt;BLOCKQUOTE&gt;
&lt;P&gt;Do all indexer machines need access to hadoop (hdfs)?&lt;BR /&gt;
Yes, they need network access, Java and Hadoop libraries (ie you should be able to successfully run hadoop fs -copyFromLocal .... from every indexer)&lt;/P&gt;

&lt;P&gt;Does hunk copy the files locally and than to hdfs?&lt;BR /&gt;
No, the buckets are copied from indexers directly to HDFS&lt;/P&gt;
&lt;/BLOCKQUOTE&gt;</description>
    <pubDate>Tue, 23 Jun 2015 19:17:16 GMT</pubDate>
    <dc:creator>Ledion_Bitincka</dc:creator>
    <dc:date>2015-06-23T19:17:16Z</dc:date>
    <item>
      <title>How archiving splunk logs to hdfs work</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172780#M34767</link>
      <description>&lt;P&gt;I know that Hunk issues the archivebuckets command, which will start the archiving process on each indexer.&lt;BR /&gt;
What is the archiving process?&lt;/P&gt;

&lt;P&gt;Do all indexer machines need access to hadoop (hdfs)?&lt;BR /&gt;
Does hunk copy the files locally and than to hdfs?&lt;/P&gt;

&lt;P&gt;please explain exactly what happens in this process.&lt;/P&gt;</description>
      <pubDate>Tue, 23 Jun 2015 13:47:29 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172780#M34767</guid>
      <dc:creator>gaddyh</dc:creator>
      <dc:date>2015-06-23T13:47:29Z</dc:date>
    </item>
    <item>
      <title>Re: How archiving splunk logs to hdfs work</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172781#M34768</link>
      <description>&lt;P&gt;&lt;A href="http://blogs.splunk.com/2015/01/21/new-in-hunk-6-2-1-splunk-archiving-searchable-archives/"&gt;The announcement  blog&lt;/A&gt; indicates all indexers need Java and Hadoop clients so I would guess that they're talking directly to HDFS.&lt;/P&gt;</description>
      <pubDate>Tue, 23 Jun 2015 14:17:49 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172781#M34768</guid>
      <dc:creator>acharlieh</dc:creator>
      <dc:date>2015-06-23T14:17:49Z</dc:date>
    </item>
    <item>
      <title>Re: How archiving splunk logs to hdfs work</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172782#M34769</link>
      <description>&lt;P&gt;Archiving indexers from Splunk to HDFS require the bucket to be in the Warm or Cold stage. Therefore, we can either wait a day, or restart Hunk.&lt;BR /&gt;
The Splunk_Archiver (New App with Hunk 6.2.1) using the Bundle Replication is distributed to all the Indexers. &amp;nbsp;&amp;nbsp;&lt;BR /&gt;
Every 60 Minutes, the App on each Indexer executes the search = | archivebuckets This command will Triggers bucket copying From the Indexer to HDFS (Similar to Hadoop fs –put BUCKET /HDFS)&lt;/P&gt;</description>
      <pubDate>Tue, 23 Jun 2015 16:01:12 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172782#M34769</guid>
      <dc:creator>rdagan_splunk</dc:creator>
      <dc:date>2015-06-23T16:01:12Z</dc:date>
    </item>
    <item>
      <title>Re: How archiving splunk logs to hdfs work</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172783#M34770</link>
      <description>&lt;BLOCKQUOTE&gt;
&lt;P&gt;Do all indexer machines need access to hadoop (hdfs)?&lt;BR /&gt;
Yes, they need network access, Java and Hadoop libraries (ie you should be able to successfully run hadoop fs -copyFromLocal .... from every indexer)&lt;/P&gt;

&lt;P&gt;Does hunk copy the files locally and than to hdfs?&lt;BR /&gt;
No, the buckets are copied from indexers directly to HDFS&lt;/P&gt;
&lt;/BLOCKQUOTE&gt;</description>
      <pubDate>Tue, 23 Jun 2015 19:17:16 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172783#M34770</guid>
      <dc:creator>Ledion_Bitincka</dc:creator>
      <dc:date>2015-06-23T19:17:16Z</dc:date>
    </item>
    <item>
      <title>Re: How archiving splunk logs to hdfs work</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172784#M34771</link>
      <description>&lt;P&gt;Which ports does Hunk use to connect to the splunk indexers?&lt;/P&gt;

&lt;P&gt;What Splunk version need to be installed on the indexer machines to support archiving indexes?&lt;BR /&gt;
What else needs to be installed on the Splunk indexer machines?&lt;/P&gt;</description>
      <pubDate>Sun, 28 Jun 2015 09:14:56 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172784#M34771</guid>
      <dc:creator>gaddyh</dc:creator>
      <dc:date>2015-06-28T09:14:56Z</dc:date>
    </item>
    <item>
      <title>Re: How archiving splunk logs to hdfs work</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172785#M34772</link>
      <description>&lt;P&gt;To connect to indexers, the same Splunk Enterprise ports are needed (8089 by default). For Hadoop, ports depend on the Hadoop distro - indexers need access to the Secondary/Namenode (usually port 8020), and DataNode transfer ports (usually port 50010). Indexer Splunk version doesn't matter, but you need Java and Hadoop libraries on the indexers. &lt;/P&gt;</description>
      <pubDate>Mon, 29 Jun 2015 17:55:02 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-archiving-splunk-logs-to-hdfs-work/m-p/172785#M34772</guid>
      <dc:creator>Ledion_Bitincka</dc:creator>
      <dc:date>2015-06-29T17:55:02Z</dc:date>
    </item>
  </channel>
</rss>

