<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic How can we find out where the delay in indexing is? in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300017#M56744</link>
    <description>&lt;P&gt;We have the following search - &lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;base search
| eval diff= _indextime - _time 
| eval capturetime=strftime(_time,"%Y-%m-%d %H:%M:%S") 
| eval indextime=strftime(_indextime,"%Y-%m-%d %H:%M:%S") 
| table capturetime indextime  diff
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;We see the following -&lt;/P&gt;

&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="alt text"&gt;&lt;img src="https://community.splunk.com/t5/image/serverpage/image-id/2958i244EFCFBF1C472C1/image-size/large?v=v2&amp;amp;px=999" role="button" title="alt text" alt="alt text" /&gt;&lt;/span&gt;&lt;/P&gt;

&lt;P&gt;So, we see a delay of over five hours in indexing. Is there a way to find out where these events "got stuck"? In this case, these events are coming from hadoop servers and the forwarder processes around 1/2 million files. We would like to know whether the delay is at the forwarder level or on the indexer side.&lt;/P&gt;</description>
    <pubDate>Wed, 17 May 2017 19:47:22 GMT</pubDate>
    <dc:creator>ddrillic</dc:creator>
    <dc:date>2017-05-17T19:47:22Z</dc:date>
    <item>
      <title>How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300017#M56744</link>
      <description>&lt;P&gt;We have the following search - &lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;base search
| eval diff= _indextime - _time 
| eval capturetime=strftime(_time,"%Y-%m-%d %H:%M:%S") 
| eval indextime=strftime(_indextime,"%Y-%m-%d %H:%M:%S") 
| table capturetime indextime  diff
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;We see the following -&lt;/P&gt;

&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="alt text"&gt;&lt;img src="https://community.splunk.com/t5/image/serverpage/image-id/2958i244EFCFBF1C472C1/image-size/large?v=v2&amp;amp;px=999" role="button" title="alt text" alt="alt text" /&gt;&lt;/span&gt;&lt;/P&gt;

&lt;P&gt;So, we see a delay of over five hours in indexing. Is there a way to find out where these events "got stuck"? In this case, these events are coming from hadoop servers and the forwarder processes around 1/2 million files. We would like to know whether the delay is at the forwarder level or on the indexer side.&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 19:47:22 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300017#M56744</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T19:47:22Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300018#M56745</link>
      <description>&lt;P&gt;Just to clarify, did you check there is no &lt;CODE&gt;maxKBps = &amp;lt;some Number other than 0&amp;gt;&lt;/CODE&gt; option set in &lt;CODE&gt;limits.conf&lt;/CODE&gt; on the UF? &lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 19:52:27 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300018#M56745</guid>
      <dc:creator>MuS</dc:creator>
      <dc:date>2017-05-17T19:52:27Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300019#M56746</link>
      <description>&lt;P&gt;ok, I see -&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;$ find . -name "limits.conf"       | xargs grep -i maxKBps
./etc/apps/universal_config_forwarder/local/limits.conf:maxKBps = 0
./etc/apps/SplunkUniversalForwarder/default/limits.conf:maxKBps = 256
./etc/system/default/limits.conf:maxKBps = 0
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Wed, 17 May 2017 19:57:52 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300019#M56746</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T19:57:52Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300020#M56747</link>
      <description>&lt;P&gt;and then - &lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;$ ./splunk btool --debug limits list | grep maxKBp
/opt/splunk/splunkforwarder/etc/apps/universal_config_forwarder/local/limits.conf maxKBps = 0
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Wed, 17 May 2017 20:01:24 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300020#M56747</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T20:01:24Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300021#M56748</link>
      <description>&lt;P&gt;I would run a btool command to check which setting is applied. (system/default has lowest priority).&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;bin/splunk btool limits list --debug | grep maxKBps
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Wed, 17 May 2017 20:03:17 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300021#M56748</guid>
      <dc:creator>somesoni2</dc:creator>
      <dc:date>2017-05-17T20:03:17Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300022#M56749</link>
      <description>&lt;P&gt;use this command to show what is actually applied as config:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt; splunk btool limits list thruput
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;that is on the forwarder. But by looks of it, you have no limit active ... Did you check &lt;CODE&gt;DMC / MC&lt;/CODE&gt; for any blocked queues?&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 20:04:00 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300022#M56749</guid>
      <dc:creator>MuS</dc:creator>
      <dc:date>2017-05-17T20:04:00Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300023#M56750</link>
      <description>&lt;P&gt;right - that's what I did...&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 20:05:10 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300023#M56750</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T20:05:10Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300024#M56751</link>
      <description>&lt;P&gt;Great. It shows -&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;$ ./splunk btool limits list thruput
[thruput]
maxKBps = 0
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Wed, 17 May 2017 20:06:03 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300024#M56751</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T20:06:03Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300025#M56752</link>
      <description>&lt;P&gt;I was late/early on that. Check the various queue sizes if there is any high spikes on the queue sizes.&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;index=_internal sourcetype=splunkd source=*metrics.log group=queue 
| timechart avg(current_size) by name
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;You can add host=yourUFName to see queue sizes on UF and host=Indexer (add more OR condition for all indexers) to see queue sizes on Indexers. You may need to adjust queue sizes based on results from there. &lt;A href="https://answers.splunk.com/answers/38218/universal-forwarder-parsingqueue-kb-size.html"&gt;https://answers.splunk.com/answers/38218/universal-forwarder-parsingqueue-kb-size.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 20:09:45 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300025#M56752</guid>
      <dc:creator>somesoni2</dc:creator>
      <dc:date>2017-05-17T20:09:45Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300026#M56753</link>
      <description>&lt;P&gt;Great. I see the following - &lt;/P&gt;

&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="alt text"&gt;&lt;img src="https://community.splunk.com/t5/image/serverpage/image-id/2955i5BF5EE8AEAA5CD36/image-size/large?v=v2&amp;amp;px=999" role="button" title="alt text" alt="alt text" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 20:26:48 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300026#M56753</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T20:26:48Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300027#M56754</link>
      <description>&lt;P&gt;The  aggQueue is where date parsing and line merging happens. This suggest that there may be in-efficient event parsing configuration setup. What is the sourcetype definition (props.conf on indexers) you've for sourcetypes involved?&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 20:41:58 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300027#M56754</guid>
      <dc:creator>somesoni2</dc:creator>
      <dc:date>2017-05-17T20:41:58Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300028#M56755</link>
      <description>&lt;P&gt;Interesting - this sourcetype doesn't show up in in &lt;CODE&gt;props.conf&lt;/CODE&gt;...&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 20:49:47 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300028#M56755</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T20:49:47Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300029#M56756</link>
      <description>&lt;P&gt;It means there is no config setup and Splunk has to figure everything out, hence the spikes. I would suggest defining an efficient line breaking and event parsing for this data and get it deployed on Indexers (would need to restart indexers). I hope you'd see lower latency/queue sizes after that. If you could share some sample raw events, we can suggest some.&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 21:05:14 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300029#M56756</guid>
      <dc:creator>somesoni2</dc:creator>
      <dc:date>2017-05-17T21:05:14Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300030#M56757</link>
      <description>&lt;P&gt;perfect - I'll work on it.&lt;/P&gt;</description>
      <pubDate>Wed, 17 May 2017 21:21:04 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300030#M56757</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-17T21:21:04Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300031#M56758</link>
      <description>&lt;P&gt;Do you see any helpful information in this Management Console dashboard?&lt;BR /&gt;
Indexing Pipeline: &lt;A href="http://docs.splunk.com/Documentation/Splunk/6.6.0/DMC/IndexingInstance"&gt;http://docs.splunk.com/Documentation/Splunk/6.6.0/DMC/IndexingInstance&lt;/A&gt;&lt;BR /&gt;
Forwarders: &lt;A href="http://docs.splunk.com/Documentation/Splunk/6.6.0/DMC/ForwardersDeployment"&gt;http://docs.splunk.com/Documentation/Splunk/6.6.0/DMC/ForwardersDeployment&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Sat, 20 May 2017 20:15:36 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300031#M56758</guid>
      <dc:creator>rdagan_splunk</dc:creator>
      <dc:date>2017-05-20T20:15:36Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300032#M56759</link>
      <description>&lt;P&gt;Hi @rdagan,&lt;/P&gt;

&lt;P&gt;We had a production change on Wednesday night. On the following day, Thursday, we saw this delay in indexing -&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;base query 
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;followed by -&lt;/P&gt;

&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="alt text"&gt;&lt;img src="https://community.splunk.com/t5/image/serverpage/image-id/2956i57B5AC98F63EC43E/image-size/large?v=v2&amp;amp;px=999" role="button" title="alt text" alt="alt text" /&gt;&lt;/span&gt;&lt;/P&gt;

&lt;P&gt;On Friday there was no delay (the right column) -&lt;/P&gt;

&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="alt text"&gt;&lt;img src="https://community.splunk.com/t5/image/serverpage/image-id/2957i4C82316E9E12C09B/image-size/large?v=v2&amp;amp;px=999" role="button" title="alt text" alt="alt text" /&gt;&lt;/span&gt;&lt;/P&gt;

&lt;P&gt;And we saw this behavior before on other production changes involving this large hadoop file systems. So, I think that it takes the forwarder hours to scan this large number of files and index the right information, a day or two later all is fine. Just checked it now and it's perfect. So, the delay's time frame is around the forwarder bounce time.&lt;/P&gt;

&lt;P&gt;The thing is - what can we improve on the forwarder to lower this delay after the bounce?&lt;/P&gt;

&lt;P&gt;On the forwarder we see -&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;$ ulimit -a
core file size          (blocks, -c) 0
data seg size           (kbytes, -d) unlimited
scheduling priority             (-e) 0
file size               (blocks, -f) unlimited
pending signals                 (-i) 1033069
max locked memory       (kbytes, -l) 64
max memory size         (kbytes, -m) unlimited
open files                      (-n) 64000
pipe size            (512 bytes, -p) 8
POSIX message queues     (bytes, -q) 819200
real-time priority              (-r) 0
stack size              (kbytes, -s) 10240
cpu time               (seconds, -t) unlimited
max user processes              (-u) 1024
virtual memory          (kbytes, -v) unlimited
file locks                      (-x) unlimited
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;And thank you @MuS and @somesoni2 for validating that nothing is fundamentally wrong with either the forwarder's configuration or the index queues... &lt;/P&gt;</description>
      <pubDate>Sat, 20 May 2017 23:20:16 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/300032#M56759</guid>
      <dc:creator>ddrillic</dc:creator>
      <dc:date>2017-05-20T23:20:16Z</dc:date>
    </item>
    <item>
      <title>Re: How can we find out where the delay in indexing is?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/505456#M86120</link>
      <description>&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 22 Jun 2020 14:52:59 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/How-can-we-find-out-where-the-delay-in-indexing-is/m-p/505456#M86120</guid>
      <dc:creator>uagraw01</dc:creator>
      <dc:date>2020-06-22T14:52:59Z</dc:date>
    </item>
  </channel>
</rss>

