<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Forwarding and Indexing Large Files in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362021#M65993</link>
    <description>&lt;P&gt;Anyone have any experience with fowarding and indexing files larger than 200mb every minute?&lt;BR /&gt;
I'm curious if there are any forwarding processes or indexing processes that need some tuning to keep track of new files being created by our app each minute.&lt;/P&gt;</description>
    <pubDate>Wed, 15 Nov 2017 00:26:55 GMT</pubDate>
    <dc:creator>rbarajas</dc:creator>
    <dc:date>2017-11-15T00:26:55Z</dc:date>
    <item>
      <title>Forwarding and Indexing Large Files</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362021#M65993</link>
      <description>&lt;P&gt;Anyone have any experience with fowarding and indexing files larger than 200mb every minute?&lt;BR /&gt;
I'm curious if there are any forwarding processes or indexing processes that need some tuning to keep track of new files being created by our app each minute.&lt;/P&gt;</description>
      <pubDate>Wed, 15 Nov 2017 00:26:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362021#M65993</guid>
      <dc:creator>rbarajas</dc:creator>
      <dc:date>2017-11-15T00:26:55Z</dc:date>
    </item>
    <item>
      <title>Re: Forwarding and Indexing Large Files</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362022#M65994</link>
      <description>&lt;P&gt;Hi rbarajas,&lt;BR /&gt;
I experienced this situation, the only way to avoid problems is to have a really good infrastructure: at first a really quick storage (1200 iops), many CPUs on Indexers and a good network, large files indexing could have long indexing queues that cause slow performaces in all the system.&lt;BR /&gt;
About network, Splunk optimize bandwidth occupation, if you haven't a great constrain you could change the network parameter and send larger packets but I usually don't do this!&lt;BR /&gt;
As second consequence you could not have a really near time monitoring because there could be a delay in indexing: this is important for real time monitoring and acceletarions. &lt;BR /&gt;
Bye.&lt;BR /&gt;
Giuseppe&lt;/P&gt;</description>
      <pubDate>Wed, 15 Nov 2017 09:55:39 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362022#M65994</guid>
      <dc:creator>gcusello</dc:creator>
      <dc:date>2017-11-15T09:55:39Z</dc:date>
    </item>
    <item>
      <title>Re: Forwarding and Indexing Large Files</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362023#M65995</link>
      <description>&lt;P&gt;Blessed by a good network and robust indexers, the universal forwarder on the endpoint is usually my bottleneck. By closely watching the local splunkd.log file you can see where the forwarder is struggling.&lt;/P&gt;

&lt;P&gt;This post ended up being a major help in a recent major log source onboarding experience I had - &lt;A href="https://answers.splunk.com/answers/38218/universal-forwarder-parsingqueue-kb-size.html"&gt;https://answers.splunk.com/answers/38218/universal-forwarder-parsingqueue-kb-size.html&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;I also learned the lesson during that onboarding process to uncompress large log files before handing off to the UF for parsing. The UF can handle some on the fly decompression/parsing, but at a certain point it can't keep things straight... We were dealing with compressed text files that were 1-3 million lines long, not a normal type of log situation though. &lt;/P&gt;

&lt;P&gt;Best,&lt;/P&gt;

&lt;P&gt;David&lt;/P&gt;</description>
      <pubDate>Wed, 15 Nov 2017 13:46:30 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Forwarding-and-Indexing-Large-Files/m-p/362023#M65995</guid>
      <dc:creator>djl</dc:creator>
      <dc:date>2017-11-15T13:46:30Z</dc:date>
    </item>
  </channel>
</rss>

