<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Does Hunk take .snappy files from Hadoop as an input? in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115468#M24138</link>
    <description>&lt;P&gt;What Hadoop version are you using?  What does the following command return:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt; jar -tf $HADOOP_HOME/hadoop-core-*.jar | grep Snappy
&lt;/CODE&gt;&lt;/PRE&gt;</description>
    <pubDate>Thu, 18 Sep 2014 20:33:09 GMT</pubDate>
    <dc:creator>Ledion_Bitincka</dc:creator>
    <dc:date>2014-09-18T20:33:09Z</dc:date>
    <item>
      <title>Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115460#M24130</link>
      <description>&lt;P&gt;Does Hunk takes .snappy files from hadoop as input...when we are trying to do so...we are getting the following error message&lt;/P&gt;

&lt;P&gt;09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -  Exception in thread "main" java.lang.UnsatisfiedLinkError: org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy()Z&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy(Native Method)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:62)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.io.compress.SnappyCodec.getDecompressorType(SnappyCodec.java:185)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.io.compress.CodecPool.getDecompressor(CodecPool.java:131)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.mapreduce.lib.input.LineRecordReader.initialize(LineRecordReader.java:91)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.SplunkLineRecordReader.vixInitialize(SplunkLineRecordReader.java:17)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.BaseSplunkRecordReader.initialize(BaseSplunkRecordReader.java:76)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.JobSubmitterInputFormat.createRecordReader(JobSubmitterInputFormat.java:64)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkBaseMapper.stream(SplunkBaseMapper.java:319)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR$SearchHandler.streamData(SplunkMR.java:604)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR$SearchHandler$1.accept(SplunkMR.java:616)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR$SearchHandler$1.accept(SplunkMR.java:613)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.FileSplitGenerator.sendSplitToAcceptor(FileSplitGenerator.java:27)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.FileSplitGenerator.generateSplits(FileSplitGenerator.java:81)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.VirtualIndex$FileSplitter.accept(VirtualIndex.java:992)&lt;BR /&gt;
09-08-2014 19:28:25.530 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.VirtualIndex$FileSplitter.accept(VirtualIndex.java:970)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.VirtualIndex$VIXPathSpecifier.addStatus(VirtualIndex.java:269)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.VirtualIndex$VIXPathSpecifier.listStatus(VirtualIndex.java:381)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.VirtualIndex.generateSplits(VirtualIndex.java:1050)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.input.VixSplitGenerator.generateSplits(VixSplitGenerator.java:55)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR$SearchHandler.streamData(SplunkMR.java:634)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR$SearchHandler.executeImpl(SplunkMR.java:850)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR$SearchHandler.execute(SplunkMR.java:695)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR.runImpl(SplunkMR.java:1295)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at com.splunk.mr.SplunkMR.run(SplunkMR.java:1087)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)&lt;BR /&gt;
09-08-2014 19:28:25.531 INFO  ERP.MyHadoopProvider -    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)&lt;/P&gt;</description>
      <pubDate>Mon, 08 Sep 2014 14:42:23 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115460#M24130</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2014-09-08T14:42:23Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115461#M24131</link>
      <description>&lt;P&gt;Yes, Hunk supports snappy and all other compressions supported by Hadoop. In your particular case however, the Hadoop libraries are having an issue with loading the native code for snappy &lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;09-08-2014 19:28:25.530 INFO ERP.MyHadoopProvider - Exception in thread "main" java.lang.UnsatisfiedLinkError: org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy()Z
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;What version of Hadoop are you using?&lt;/P&gt;

&lt;P&gt;Are the snappy native libs in your: $HADOOP_HOME/lib/native ?&lt;/P&gt;</description>
      <pubDate>Mon, 08 Sep 2014 19:55:14 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115461#M24131</guid>
      <dc:creator>Ledion_Bitincka</dc:creator>
      <dc:date>2014-09-08T19:55:14Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115462#M24132</link>
      <description>&lt;P&gt;Yes native libs are in our  $HADOOP_HOME/lib/native  and in our hadoop env "hadoop fs -text *.snappy" is working fine. We are using hadoop-2.0.0-cdh4.7.0 version.&lt;/P&gt;</description>
      <pubDate>Tue, 09 Sep 2014 05:40:22 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115462#M24132</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2014-09-09T05:40:22Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115463#M24133</link>
      <description>&lt;P&gt;This issue is solved as we moved snappy native libs to $HADOOP_HOME/lib/native. But now we are getting the error&lt;/P&gt;

&lt;P&gt;09-09-2014 12:21:50.492 INFO  ERP.Hadoop_Provider -  Exception in thread "main" java.lang.OutOfMemoryError: Java heap space&lt;BR /&gt;
09-09-2014 12:21:50.493 INFO  ERP.Hadoop_Provider -      at org.apache.hadoop.io.compress.BlockDecompressorStream.getCompressedData(BlockDecompressorStream.java:115)&lt;BR /&gt;
09-09-2014 12:21:50.493 INFO  ERP.Hadoop_Provider -      at org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockDecompressorStream.java:95)&lt;/P&gt;</description>
      <pubDate>Mon, 28 Sep 2020 17:31:42 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115463#M24133</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2020-09-28T17:31:42Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115464#M24134</link>
      <description>&lt;P&gt;We have tried to increase hadoop heap size and java heap size to almost 4 to 5 GB issue still remains the same&lt;/P&gt;</description>
      <pubDate>Tue, 09 Sep 2014 06:59:44 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115464#M24134</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2014-09-09T06:59:44Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115465#M24135</link>
      <description>&lt;P&gt;This issue is also resolved...now getting the following error&lt;/P&gt;

&lt;P&gt;09-09-2014 12:46:34.704 ERROR ERP.Hadoop_Provider -  SplunkMR$SearchHandler$1 - Unexpected end of block in input stream&lt;BR /&gt;
09-09-2014 12:46:34.704 ERROR ERP.Hadoop_Provider -  java.io.EOFException: Unexpected end of block in input stream&lt;BR /&gt;
09-09-2014 12:46:34.704 ERROR ERP.Hadoop_Provider -      at org.apache.hadoop.io.compress.BlockDecompressorStream.getCompressedData(BlockDecompressorStream.java:121)&lt;/P&gt;</description>
      <pubDate>Mon, 28 Sep 2020 17:31:45 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115465#M24135</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2020-09-28T17:31:45Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115466#M24136</link>
      <description>&lt;P&gt;I'm sure you've checked this, but double checking - are you sure the file is not corrupt? ie can the file be fully read by hadoop fs -text ...&lt;/P&gt;</description>
      <pubDate>Tue, 09 Sep 2014 16:38:42 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115466#M24136</guid>
      <dc:creator>Ledion_Bitincka</dc:creator>
      <dc:date>2014-09-09T16:38:42Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115467#M24137</link>
      <description>&lt;P&gt;We are still facing with snappy conversion....actually when we got the Block decompressor error...we added the following property in core-site.xml to add snappycodec class &lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;&amp;lt;name&amp;gt;io.compression.codecs&amp;lt;/name&amp;gt;
&amp;lt;value&amp;gt;
  org.apache.hadoop.io.compress.SnappyCodec
&amp;lt;/value&amp;gt;
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;and now we are getting the below error now....the error is saying org.apache.hadoop.io.compress.SnappyCodec class not found...but we have the actuall hadoop-commnsxxx.jar in the classpath...not sure why we are getting this error...Any help on this is greatly appreciated...&lt;/P&gt;

&lt;P&gt;09-18-2014 20:42:58.184 INFO  ERP.Hadoop_Provider -  ClusterInfoLogger - Hadoop cluster spec: provider=Hadoop_Provider, tasktrackers=2, map_inuse=1, map_slots=20, reduce_inuse=1, reduce_slots=4&lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -  SplunkMR - Compression codec &lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -        org.apache.hadoop.io.compress.SnappyCodec&lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -       not found.&lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -  java.lang.IllegalArgumentException: Compression codec &lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -        org.apache.hadoop.io.compress.SnappyCodec&lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -       not found.&lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -     at org.apache.hadoop.io.compress.CompressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:134)&lt;BR /&gt;
09-18-2014 20:42:58.246 ERROR ERP.Hadoop_Provider -     at org.apache.hadoop.io.compress.CompressionCodecFactory.&lt;INIT&gt;(CompressionCodecFactory.java:174)&lt;/INIT&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 28 Sep 2020 17:37:15 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115467#M24137</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2020-09-28T17:37:15Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115468#M24138</link>
      <description>&lt;P&gt;What Hadoop version are you using?  What does the following command return:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt; jar -tf $HADOOP_HOME/hadoop-core-*.jar | grep Snappy
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Thu, 18 Sep 2014 20:33:09 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115468#M24138</guid>
      <dc:creator>Ledion_Bitincka</dc:creator>
      <dc:date>2014-09-18T20:33:09Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115469#M24139</link>
      <description>&lt;P&gt;We are using Hadoop version as: Hadoop 2.0.0-cdh4.7.0&lt;/P&gt;

&lt;P&gt;We are using hadoop yarn based mapreduce, There is no hadoop-core-&lt;EM&gt;.jar in our Cdh distribution but we found Snappy in hadoop-common-&lt;/EM&gt;.jar. Below are the results...&lt;/P&gt;

&lt;P&gt;$  jar -tf $HADOOP_HOME/lib/hadoop-common-*.jar | grep Snappy&lt;/P&gt;

&lt;P&gt;org/apache/hadoop/io/compress/SnappyCodec.class&lt;BR /&gt;
org/apache/hadoop/io/compress/snappy/SnappyCompressor.class&lt;BR /&gt;
org/apache/hadoop/io/compress/snappy/SnappyDecompressor.class&lt;/P&gt;</description>
      <pubDate>Fri, 19 Sep 2014 06:15:57 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115469#M24139</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2014-09-19T06:15:57Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115470#M24140</link>
      <description>&lt;P&gt;What version of Hunk are you using?&lt;/P&gt;

&lt;P&gt;This definitely seems like a classpath issue - what is your HADOOP_HOME pointing to? I just downloaded CDH4.7.0 and snappy seems to be in HADOOP_HOME/share/hadoop/common/hadoop-common-2.0.0-cdh4.7.0.jar&lt;/P&gt;</description>
      <pubDate>Mon, 28 Sep 2020 17:38:26 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115470#M24140</guid>
      <dc:creator>Ledion_Bitincka</dc:creator>
      <dc:date>2020-09-28T17:38:26Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115471#M24141</link>
      <description>&lt;P&gt;We are using hunk 6.1.2 version. &lt;BR /&gt;
HADOOP_HOME is pointing to installation directory of Hadoop. and as you mentioned, snappy is included in hadoop-common-*.jar file only. We have also copied the same jar into lib directory of hadoop installation. So, could you please suggest how to setup the classpath for accessing this snappy codec during run time of hunk.&lt;/P&gt;</description>
      <pubDate>Mon, 22 Sep 2014 05:25:42 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115471#M24141</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2014-09-22T05:25:42Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115472#M24142</link>
      <description>&lt;P&gt;HI &lt;/P&gt;

&lt;P&gt;We have still not resolved the issue...please help us in resolving this snappy connectivity issue....&lt;/P&gt;</description>
      <pubDate>Tue, 14 Oct 2014 11:46:56 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115472#M24142</guid>
      <dc:creator>balagovardhan</dc:creator>
      <dc:date>2014-10-14T11:46:56Z</dc:date>
    </item>
    <item>
      <title>Re: Does Hunk take .snappy files from Hadoop as an input?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115473#M24143</link>
      <description>&lt;P&gt;Is there a way for us to test one of your snappy files? Do you have an email we can connect?&lt;/P&gt;</description>
      <pubDate>Tue, 14 Oct 2014 16:21:50 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Does-Hunk-take-snappy-files-from-Hadoop-as-an-input/m-p/115473#M24143</guid>
      <dc:creator>rdagan_splunk</dc:creator>
      <dc:date>2014-10-14T16:21:50Z</dc:date>
    </item>
  </channel>
</rss>

