<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines? in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610008#M105727</link>
    <description>&lt;P&gt;We are already fixing this issue where a remote node(UF/HF) can pass down shutdown key to next layer. Upcoming set of patches(9.0.2 etc) has the fix.&amp;nbsp; However with this diag we can verify if it's same issue .&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 18 Aug 2022 16:35:02 GMT</pubDate>
    <dc:creator>hrawat</dc:creator>
    <dc:date>2022-08-18T16:35:02Z</dc:date>
    <item>
      <title>Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/604320#M105104</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;
&lt;P&gt;We've recently tested out a new path for data to flow into our Splunk environment from Universal Forwarders.&lt;BR /&gt;We have rolled this out to a small porsion of our UFs for now, and we've been running into an issue a few times already, that I can't figure out.&lt;/P&gt;
&lt;P&gt;Data flow is as follows:&lt;BR /&gt;UFs sending logs with [httpout] -&amp;gt; Load Balancer -&amp;gt; A set of Heavy Forwarders receiving data via [http://..], and forwarding data with [tcpout] -&amp;gt; Index cluster receiving data with [tcp://..].&lt;/P&gt;
&lt;P&gt;The heavy forwarders are there to do some filtering, and also routing of specific data to other Splunk environments. The Heavy Forwarders are also configured with&amp;nbsp;parallelIngestionPipelines=2.&lt;BR /&gt;I can also mention that all parts of this environment is running on Windows.&lt;/P&gt;
&lt;P&gt;The issue:&lt;BR /&gt;A few times I've suddenly had 100% queue on all the four data queues(parsing/aggregation/typing/indexing) on one of the Heavy Forwarders and only on one of its pipelines, but not the other. It turns out after looking in splunkd.log, that it seems like one pipeline has simply shut down (I didn't know this was even possible).&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;At first, splunkd.log is filled with the usual over and over:&lt;BR /&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;LI-CODE lang="markup"&gt;07-01-2022 12:35:55.724 +0200 INFO  TailReader [2092 tailreader1] - Batch input finished reading file='D:\Splunk\var\spool\splunk\tracker.log'
07-01-2022 12:35:59.129 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Connected to idx=X.X.X.X:9997:0, pset=1, reuse=0. autoBatch=1
07-01-2022 12:35:59.602 +0200 INFO  AutoLoadBalancedConnectionStrategy [7160 TcpOutEloop] - Connected to idx=Y.Y.Y.Y:9997:0, pset=0, reuse=0. autoBatch=1&lt;/LI-CODE&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Until it suddenly logs this one or a few times:&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;LI-CODE lang="markup"&gt;07-01-2022 12:36:15.628 +0200 WARN  TcpOutputProc [4360 indexerPipe_1] - Pipeline data does not have indexKey. [_path] = C:\Program Files\SplunkUniversalForwarder\bin\splunk-winevtlog.exe\n[_raw] = \n[_meta] = punct::\n[_stmid] = FSDNhUnXitGgjKJ.C\n[MetaData:Source] = source::WinEventLog\n[MetaData:Host] = host::HOSTNAME-1\n[MetaData:Sourcetype] = sourcetype::WinEventLog\n[_done] = _done\n[_linebreaker] = _linebreaker\n[_conf] = source::WinEventLog|host::HOSTNAME-1|WinEventLog|\n
07-01-2022 12:36:18.966 +0200 WARN  TcpOutputProc [4360 indexerPipe_1] - Pipeline data does not have indexKey. [_path] = C:\Program Files\SplunkUniversalForwarder\bin\splunk-winevtlog.exe\n[_raw] = \n[_meta] = punct::\n[_stmid] = WpC6rhVY9yDblPB.C\n[MetaData:Source] = source::WinEventLog\n[MetaData:Host] = host::HOSTNAME-2\n[MetaData:Sourcetype] = sourcetype::WinEventLog\n[_done] = _done\n[_linebreaker] = _linebreaker\n[_conf] = source::WinEventLog|host::HOSTNAME-2|WinEventLog|\n&lt;/LI-CODE&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Then in the same millisecond this happens:&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;LI-CODE lang="markup"&gt;07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Shutting down auto load balanced connection strategy
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Auto load balanced connection strategy shutdown finished
07-01-2022 12:36:18.966 +0200 INFO  TcpOutputProc [4360 indexerPipe_1] - Waiting for TcpOutputGroups to shutdown
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Shutting down auto load balanced connection strategy
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Auto load balanced connection strategy shutdown finished
07-01-2022 12:36:19.968 +0200 INFO  TcpOutputProc [4360 indexerPipe_1] - Received shutdown control key.
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - shutting down: start
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_audit Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_configtracker Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_internal Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_thefishbucket Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_introspection Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics_rollup Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_telemetry Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=history Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=main Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=summary Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - request state change from=RUN to=SHUTDOWN_IN_PROGRESS
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_audit Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_configtracker Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_internal Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_introspection Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics_rollup Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_telemetry Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_thefishbucket Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=history Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=main Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=summary Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_metrics
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_audit
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_configtracker
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_internal
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_thefishbucket
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_introspection
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_metrics_rollup
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_telemetry
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=history
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=main
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=summary
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - request state change from=SHUTDOWN_IN_PROGRESS to=SHUTDOWN_COMPLETE
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - shutting down: end&lt;/LI-CODE&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;So it seems like pipeline-1 shut down, and pipeline-0 continues to work(this over and over, with only pset=0):&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;LI-CODE lang="markup"&gt;07-01-2022 12:36:29.527 +0200 INFO  AutoLoadBalancedConnectionStrategy [7160 TcpOutEloop] - Connected to idx=X.X.X.X:9997:0, pset=0, reuse=0. autoBatch=1&lt;/LI-CODE&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Pipeline-1 seems to still receive data somehow, as I can see queues for that one pipeline is constantly 100% in metrics.log, and we're also loosing logs to our indexers.&lt;BR /&gt;And then this starts appearing in splunkd.log over and over until I restart splunk:&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;LI-CODE lang="markup"&gt;07-01-2022 12:36:30.522 +0200 ERROR PipelineComponent [8012 CallbackRunnerThread] - Monotonic time source didn't increase; is it stuck?
07-01-2022 12:36:32.525 +0200 ERROR PipelineComponent [8012 CallbackRunnerThread] - Monotonic time source didn't increase; is it stuck?
07-01-2022 12:36:34.528 +0200 ERROR PipelineComponent [8012 CallbackRunnerThread] - Monotonic time source didn't increase; is it stuck?
07-01-2022 12:36:36.532 +0200 ERROR PipelineComponent [8012 CallbackRunnerThread] - Monotonic time source didn't increase; is it stuck?&lt;/LI-CODE&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I've had this happen to me at least 3-4 times the last week. The heavy forwarders are running Splunk 9.0.0, but I also had this happen once a few weeks ago when I ran 8.2.4. I didn't think too much of it at the time, as I was in the process of replacing them with new servers anyway, and with 9.0.0. But the logs and the symptoms are exactly the same on 9.0.0 as they were that one time at 8.2.4.&lt;/P&gt;
&lt;P&gt;I tried looking into the "Pipeline data does not have indexKey"-part, but as this is coming from UFs with identical config where all input stanzas has defined an index, I couldn't find anything wrong there.&lt;BR /&gt;That said, even if I were missing index for some log sources, the Heavy Forwarders should not just randomly shut down a pipeline anyway.&lt;/P&gt;
&lt;P&gt;Has anyone seen this happen before, or have any suggestion to what I should try or check.&lt;BR /&gt;It really seems like a bug to me, but I of course don't know that.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Best Regards&lt;BR /&gt;Morten&lt;/P&gt;</description>
      <pubDate>Tue, 12 Jul 2022 13:31:58 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/604320#M105104</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-07-12T13:31:58Z</dc:date>
    </item>
    <item>
      <title>Re: Heavy Forwarders randomly shuts down one of its parallel pipelines</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605223#M105209</link>
      <description>&lt;P&gt;Whenever you see something like&lt;/P&gt;&lt;PRE&gt; Pipeline data does not have indexKey. [_path] = C:\Program Files\SplunkUniversalForwarder\bin\splunk-winevtlog.exe\n[_raw] &lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The log contains the executable/script name. It simply means that script/exe in the log crashed.&amp;nbsp;&lt;BR /&gt;As of now it's known issue for the exec in question.&lt;BR /&gt;&lt;A href="https://community.splunk.com/t5/Getting-Data-In/Why-is-splunk-winevtlog-exe-crash-low-thruput-high-cpu-and-other/td-p/605180" target="_blank" rel="noopener"&gt;https://community.splunk.com/t5/Getting-Data-In/Why-is-splunk-winevtlog-exe-crash-low-thruput-high-cpu-and-other/td-p/605180&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;Regarding blocked queues, did you restart splunk ?&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;request state change from=SHUTDOWN_IN_PROGRESS to=SHUTDOWN_COMPLETE&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;Or TcpOut reloaded itself?&lt;BR /&gt;One of the reason for TcpOut reload is, if outputs.conf get's modified either by DS or call to tcpout rest endpoint.&lt;/P&gt;&lt;P&gt;Did you see any log "&lt;SPAN&gt;WARN DC:DeploymentClient - Restarting Splunkd" ?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 12 Jul 2022 02:51:31 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605223#M105209</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-07-12T02:51:31Z</dc:date>
    </item>
    <item>
      <title>Re: Heavy Forwarders randomly shuts down one of its parallel pipelines</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605232#M105213</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;Thank you for the reply!&lt;BR /&gt;&lt;BR /&gt;Glad to know there is a known issue with &lt;SPAN&gt;splunk-winevtlog.exe&amp;nbsp;&lt;/SPAN&gt;that will be fixed in 9.0.1.&lt;BR /&gt;I assume this bug also affects Universal Forwarders, as it's pretty much the same executable?&lt;BR /&gt;We've had UFs stop sending windows event log many times over the last year, and I've noticed&amp;nbsp;&lt;SPAN&gt;splunk-winevtlog.exe had stopped running, but I never got around to report it as it was so sporadic.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Regarding the blocked queue messages and the shutdown message, I did eventually restart splunk on the heavy forwarder in question, but that shutdown message you quoted came long before that.&lt;BR /&gt;There are no entries matching '&lt;SPAN&gt;DeploymentClient' at all in splunkd.log leading up to this.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;What happened is, right after the "pipeline data does not have indexKey" entry in the log, on the exact same timestamp, one of the pipelines(in lack of a better way of explaining it) shut down immediately.&lt;BR /&gt;&lt;SPAN&gt;&lt;BR /&gt;I realized I split up the different parts of the log in my previous post, so I've included below the log in its entirety from start of the error to shutdown complete.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;Splunk continues to run, and the other pipeline continues to forward data as expected, but the "crashed pipeline" got 100% on all four queues and were not doing anything from what I could tell, but I don't know how to verify that.&lt;BR /&gt;I didn't always have two pipelines by the way, and when I ran with just one pipeline, the same thing happened and the same log messages were logged. Splunk continued to run, but then the single and only pipeline had 100% queues.&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;07-01-2022 12:36:18.966 +0200 WARN  TcpOutputProc [4360 indexerPipe_1] - Pipeline data does not have indexKey. [_path] = C:\Program Files\SplunkUniversalForwarder\bin\splunk-winevtlog.exe\n[_raw] = \n[_meta] = punct::\n[_stmid] = WpC6rhVY9yDblPB.C\n[MetaData:Source] = source::WinEventLog\n[MetaData:Host] = host::HOSTNAME-1\n[MetaData:Sourcetype] = sourcetype::WinEventLog\n[_done] = _done\n[_linebreaker] = _linebreaker\n[_conf] = source::WinEventLog|host::HOSTNAME-1|WinEventLog|\n
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Shutting down auto load balanced connection strategy
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Auto load balanced connection strategy shutdown finished
07-01-2022 12:36:18.966 +0200 INFO  TcpOutputProc [4360 indexerPipe_1] - Waiting for TcpOutputGroups to shutdown
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Shutting down auto load balanced connection strategy
07-01-2022 12:36:18.966 +0200 INFO  AutoLoadBalancedConnectionStrategy [4860 TcpOutEloop] - Auto load balanced connection strategy shutdown finished
07-01-2022 12:36:19.968 +0200 INFO  TcpOutputProc [4360 indexerPipe_1] - Received shutdown control key.
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - shutting down: start
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_audit Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_configtracker Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_internal Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_thefishbucket Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_introspection Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics_rollup Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_telemetry Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=history Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=main Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=summary Sync before shutdown
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - request state change from=RUN to=SHUTDOWN_IN_PROGRESS
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_audit Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_configtracker Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_internal Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_introspection Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_metrics_rollup Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_telemetry Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=_thefishbucket Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=history Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=main Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  IndexWriter [4360 indexerPipe_1] - idx=summary Handling shutdown or signal, reason=1
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_metrics
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_audit
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_configtracker
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_internal
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_thefishbucket
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_introspection
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_metrics_rollup
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=_telemetry
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=history
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=main
07-01-2022 12:36:19.968 +0200 INFO  HotDBManager [4360 indexerPipe_1] - closing hot mgr for idx=summary
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - request state change from=SHUTDOWN_IN_PROGRESS to=SHUTDOWN_COMPLETE
07-01-2022 12:36:19.968 +0200 INFO  IndexProcessor [4360 indexerPipe_1] - shutting down: end
07-01-2022 12:36:23.531 +0200 ERROR PipelineComponent [8012 CallbackRunnerThread] - Monotonic time source didn't increase; is it stuck?&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 12 Jul 2022 06:09:52 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605232#M105213</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-07-12T06:09:52Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605355#M105236</link>
      <description>&lt;P&gt;You will have to check for any &lt;STRONG&gt;_reload&lt;/STRONG&gt; in splunkd_access.log just before&lt;/P&gt;&lt;PRE&gt;07-01-2022 12:36:18.966&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;BR /&gt;Without a reload or restart not possible for thread to shutdown.&lt;/P&gt;&lt;P&gt;Check logs before indexkey not found log.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jul 2022 01:37:00 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605355#M105236</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-07-13T01:37:00Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605377#M105237</link>
      <description>&lt;P&gt;I had this happen again yesterday, and there is no mention of _reload anywhere in&amp;nbsp;&lt;SPAN&gt;splunkd_access.log before the timestamp where one of the pipelines shut down.&lt;BR /&gt;This time I didn't have the indexkey not found at the same timestamp as the shutdown, it was logged about 20 seconds before. But again, this indexkey not found warning is something that is logged every now and then without any pipelines shutting down also, so it might be two different issues?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;I had the shutdown of one pipeline start at 14:21:19.265. Not pasting in the whole thing here, as it's exactly the same log events as before. This is how it started, X.X.X.X would be one of my indexers, and I have four pipelines now, that's why you see four of those connections (I'm trying around with different settings):&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;07-12-2022 14:20:48.157 +0200 WARN  TcpOutputProc [1744 indexerPipe_1] - Pipeline data does not have indexKey. [_path] = C:\Program Files\SplunkUniversalForwarder\bin\splunk-winevtlog.exe\n[_raw] = \n[_meta] = punct::\n[_stmid] = Hb4RKZfh8VK1AFB.C\n[MetaData:Source] = source::WinEventLog\n[MetaData:Host] = host::HOSTNAME-1\n[MetaData:Sourcetype] = sourcetype::WinEventLog\n[_done] = _done\n[_linebreaker] = _linebreaker\n[_conf] = source::WinEventLog|host::HOSTNAME-1|WinEventLog|\n
07-12-2022 14:21:02.533 +0200 INFO  ExecProcessor [2208 ExecProcessorSchedulerThread] - message from "D:\Splunk\bin\Python3.exe D:\Splunk\etc\apps\splunk_assist\bin\instance_id_modular_input.py" [assist::instance_id_modular_input.py:228] [get_server_roles] [8064] Fetched server roles, roles=['deployment_client', 'search_peer', 'kv_store']
07-12-2022 14:21:02.533 +0200 INFO  ExecProcessor [2208 ExecProcessorSchedulerThread] - message from "D:\Splunk\bin\Python3.exe D:\Splunk\etc\apps\splunk_assist\bin\instance_id_modular_input.py" [assist::instance_id_modular_input.py:256] [get_cluster_mode] [8064] Fetched cluster mode, mode=disabled
07-12-2022 14:21:02.533 +0200 INFO  ExecProcessor [2208 ExecProcessorSchedulerThread] - message from "D:\Splunk\bin\Python3.exe D:\Splunk\etc\apps\splunk_assist\bin\instance_id_modular_input.py" [assist::instance_id_modular_input.py:30] [should_run] [8064] should run test, sh=False
07-12-2022 14:21:07.750 +0200 INFO  TailReader [6328 tailreader1] - Batch input finished reading file='D:\Splunk\var\spool\splunk\tracker.log'
07-12-2022 14:21:15.004 +0200 INFO  AutoLoadBalancedConnectionStrategy [5784 TcpOutEloop] - Found currently active indexer. Connected to idx=X.X.X.X:9998:0, reuse=1.
07-12-2022 14:21:15.074 +0200 INFO  AutoLoadBalancedConnectionStrategy [7316 TcpOutEloop] - Found currently active indexer. Connected to idx=X.X.X.X:9998:0, reuse=1.
07-12-2022 14:21:16.073 +0200 INFO  AutoLoadBalancedConnectionStrategy [5668 TcpOutEloop] - Found currently active indexer. Connected to idx=X.X.X.X:9998:0, reuse=1.
07-12-2022 14:21:16.198 +0200 INFO  AutoLoadBalancedConnectionStrategy [4084 TcpOutEloop] - Found currently active indexer. Connected to idx=X.X.X.X:9998:0, reuse=1.
07-12-2022 14:21:17.828 +0200 INFO  ExecProcessor [2208 ExecProcessorSchedulerThread] - message from "D:\Splunk\bin\Python3.exe D:\Splunk\etc\apps\splunk_assist\bin\instance_id_modular_input.py" [assist::instance_id_modular_input.py:228] [get_server_roles] [4608] Fetched server roles, roles=['deployment_client', 'search_peer', 'kv_store']
07-12-2022 14:21:17.828 +0200 INFO  ExecProcessor [2208 ExecProcessorSchedulerThread] - message from "D:\Splunk\bin\Python3.exe D:\Splunk\etc\apps\splunk_assist\bin\instance_id_modular_input.py" [assist::instance_id_modular_input.py:256] [get_cluster_mode] [4608] Fetched cluster mode, mode=disabled
07-12-2022 14:21:19.264 +0200 INFO  AutoLoadBalancedConnectionStrategy [5784 TcpOutEloop] - Shutting down auto load balanced connection strategy
07-12-2022 14:21:19.264 +0200 INFO  AutoLoadBalancedConnectionStrategy [5784 TcpOutEloop] - Auto load balanced connection strategy shutdown finished
07-12-2022 14:21:19.264 +0200 INFO  AutoLoadBalancedConnectionStrategy [5784 TcpOutEloop] - Shutting down auto load balanced connection strategy
07-12-2022 14:21:19.265 +0200 INFO  AutoLoadBalancedConnectionStrategy [5784 TcpOutEloop] - Auto load balanced connection strategy shutdown finished
07-12-2022 14:21:19.265 +0200 INFO  TcpOutputProc [6160 indexerPipe_0] - Received shutdown control key.
07-12-2022 14:21:19.265 +0200 INFO  IndexProcessor [6160 indexerPipe_0] - shutting down: start
.....
Rest of the shut down of pipeline 0, identical as posted before&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Here is what's in splunkd_access.log around the same time. I have masked a few thing, 1.2.3.4 would be my monitoring console for instance:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:02.542 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:02.569 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.739 +0200] "GET /services/properties/telemetry/general/sendSupportUsage HTTP/1.0" 200 5 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.751 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.763 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.787 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.799 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.809 +0200] "GET /services/properties/telemetry/general/sendSupportUsage HTTP/1.0" 200 5 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.821 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:17.833 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:18.079 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:18.091 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:18.103 +0200] "GET //servicesNS/nobody/splunk_secure_gateway/storage/collections/data/meta/soc2_opt_in HTTP/1.0" 404 140 "-" "Python-httplib2/0.13.1 (gzip)" - - - 2ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:18.116 +0200] "GET //services/server/info?output_mode=json HTTP/1.0" 200 2074 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:18.912 +0200] "GET /services/server/info HTTP/1.1" 200 5935 "-" "curl" - - - 1ms
1.2.3.4 - - [12/Jul/2022:14:21:23.318 +0200] "POST /services/admin/auth-tokens HTTP/1.1" 201 871 "-" "Splunk/9.0.0 (Windows Server 10 Standard Edition; arch=x64)" - - - 1ms
1.2.3.4 - splunk-system-user [12/Jul/2022:14:21:23.323 +0200] "GET /services/server/info HTTP/1.1" 200 1477 "-" "Splunk/9.0.0 (Windows Server 10 Standard Edition; arch=x64)" - - - 1ms
1.2.3.4 - - [12/Jul/2022:14:21:23.344 +0200] "POST /services/admin/auth-tokens HTTP/1.1" 201 871 "-" "Splunk/9.0.0 (Windows Server 10 Standard Edition; arch=x64)" - - - 2ms
1.2.3.4 - splunk-system-user [12/Jul/2022:14:21:23.355 +0200] "GET /services/admin/bundles/SplunkMonitoringConsole?count=-1 HTTP/1.1" 200 1068 "-" "Splunk/9.0.0 (Windows Server 10 Standard Edition; arch=x64)" - - - 1ms
1.2.3.4 - splunk-system-user [12/Jul/2022:14:21:23.363 +0200] "GET /services/data/indexes?mode=minimum&amp;amp;count=0&amp;amp;datatype=all HTTP/1.1" 200 1351 "-" "Splunk/9.0.0 (Windows Server 10 Standard Edition; arch=x64)" - - - 1ms
1.2.3.4 - splunk-system-user [12/Jul/2022:14:21:23.370 +0200] "GET /services/messages?count=1000&amp;amp;search=name%21%3Dremote:* HTTP/1.1" 200 493 "-" "Splunk/9.0.0 (Windows Server 10 Standard Edition; arch=x64)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.501 +0200] "GET /services/properties/telemetry/general/sendSupportUsage HTTP/1.0" 200 5 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.513 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.539 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.544 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.565 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.600 +0200] "GET /services/properties/telemetry/general/sendSupportUsage HTTP/1.0" 200 5 "-" "Python-httplib2/0.13.1 (gzip)" - - - 0ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.611 +0200] "GET //services/server/roles?output_mode=json HTTP/1.0" 200 787 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms
127.0.0.1 - splunk-system-user [12/Jul/2022:14:21:32.633 +0200] "GET //services/cluster/config?output_mode=json HTTP/1.0" 200 2610 "-" "Python-httplib2/0.13.1 (gzip)" - - - 1ms&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I went over several of these shut downs, and there is no mention of _reload in splunkd_access.log around or before any of those occurrences.&lt;BR /&gt;I guess even if it was, it should not just reload itself and/or shut down, but I guess it would have been a clue as to what is going on.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jul 2022 05:36:11 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605377#M105237</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-07-13T05:36:11Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605386#M105239</link>
      <description>&lt;P&gt;I have two heavy forwarders running behind load balancer, and I currently have four parallel pipelines on each.&lt;BR /&gt;I just now had three pipelines shut down on one of them and two pipelines shut down on the other one, within minutes of each other, right as people starts logging on to work. So it might be either load dependent, or just in general a higher chance for this to happen with more logs flowing through I guess.&lt;BR /&gt;&lt;BR /&gt;For now I'm going to move the UFs back to the old solution. where they would use tcpout directly to the indexers. Not ideal for our needs, but that would have to work for now.&lt;/P&gt;&lt;P&gt;I should probably open a support ticket about this (after my four weeks of vacation, which starts in a few days), if you or anyone else doesn't have any more idea what to test or check here.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jul 2022 06:19:04 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605386#M105239</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-07-13T06:19:04Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605829#M105303</link>
      <description>&lt;PRE&gt;Received shutdown control key.&lt;/PRE&gt;&lt;P&gt;This log is an indication of splunkd service getting shutdown signal.&lt;BR /&gt;But if you see no other processors shutting down around that time, that's definitely unexpected. If you have time, can you find if any other UF was shutting down few seconds before above log was seen on HF?&lt;/P&gt;</description>
      <pubDate>Fri, 15 Jul 2022 17:21:42 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/605829#M105303</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-07-15T17:21:42Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609574#M105680</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;Thanks for your reply, and sorry for the delayed feedback, I've been on vacation.&lt;/P&gt;&lt;P&gt;Finding if any UF was shutting down at the same time or right before will pretty much be impossible for us to do, as we manage 25000+ UFs through this solution and a lot of them are client computers which I guess shut down and restart all the time, depending on what users do.&lt;BR /&gt;We also had to stop ingesting intenal logs from UFs at some point because of the huge amount of data this produces, and we only really turn on ingestion of internal logs for specific sets of UFs if we need to troubleshoot anything.&lt;/P&gt;&lt;P&gt;I had to roll back this whole setup to our old setup before I went on vacation, to have a working solution while I was gone.&amp;nbsp;So currently this solution is not in use. But this is something we really need to get working properly, so I have to look further into this.&lt;/P&gt;</description>
      <pubDate>Tue, 16 Aug 2022 06:20:19 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609574#M105680</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-08-16T06:20:19Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609643#M105691</link>
      <description>&lt;P&gt;Do you mind sharing splunkd.log?&lt;/P&gt;</description>
      <pubDate>Tue, 16 Aug 2022 13:54:10 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609643#M105691</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-08-16T13:54:10Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609760#M105697</link>
      <description>&lt;P&gt;Do you mean the whole file? I would then want to mask quite a lot in there before posting it here, if I'm not going through a support case.&lt;BR /&gt;I did post the "relevant" parts of splunkd.log previously here, but I get how it's better to see the whole picture.&lt;/P&gt;&lt;P&gt;Currently I've rolled back to out old solution, so these heavy forwarders are not really in use and haven't "crashed" in a while.&lt;/P&gt;&lt;P&gt;I guess you would need to look at splunkd.log while also one of these crashes occurs?&lt;/P&gt;&lt;P&gt;Do you think this issue can be relevant to the other thread you posted a link to, that is fixed in 9.0.1? There is also several "indexer crashes" due to time format config issues that is fixed in 9.0.1. I just saw the release drop.&lt;BR /&gt;I might as well upgrade to 9.0.1 just because.&lt;/P&gt;</description>
      <pubDate>Wed, 17 Aug 2022 06:16:33 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609760#M105697</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-08-17T06:16:33Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609825#M105701</link>
      <description>&lt;P&gt;&amp;gt;&lt;SPAN&gt;I guess you would need to look at splunkd.log while also one of these crashes occurs?&lt;BR /&gt;&lt;/SPAN&gt;Right.&amp;nbsp;&lt;BR /&gt;But I would say open a splunk support case attach diag and let me know the case number.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 17 Aug 2022 15:00:34 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609825#M105701</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-08-17T15:00:34Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609826#M105702</link>
      <description>&lt;P&gt;&amp;gt;&lt;SPAN&gt;Do you think this issue can be relevant to the other thread you posted a link to, that is fixed in 9.0.1?&lt;BR /&gt;&lt;/SPAN&gt;No it's something unrelated to that. Chances are still an issue with 9.0.1.&lt;/P&gt;</description>
      <pubDate>Wed, 17 Aug 2022 15:02:26 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609826#M105702</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-08-17T15:02:26Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609911#M105716</link>
      <description>&lt;P&gt;Thanks, I'll put some load on this setup again and open a support case when I have another crash. Will post the case number here.&lt;/P&gt;</description>
      <pubDate>Thu, 18 Aug 2022 05:36:10 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609911#M105716</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-08-18T05:36:10Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609958#M105725</link>
      <description>&lt;P&gt;Issue reproduced.&lt;/P&gt;&lt;P&gt;Ticket has been submitted:&amp;nbsp;&lt;SPAN&gt;3035621&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 18 Aug 2022 10:17:03 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/609958#M105725</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-08-18T10:17:03Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610008#M105727</link>
      <description>&lt;P&gt;We are already fixing this issue where a remote node(UF/HF) can pass down shutdown key to next layer. Upcoming set of patches(9.0.2 etc) has the fix.&amp;nbsp; However with this diag we can verify if it's same issue .&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 18 Aug 2022 16:35:02 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610008#M105727</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-08-18T16:35:02Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610009#M105728</link>
      <description>&lt;P&gt;That's great news, thank you!&lt;/P&gt;&lt;P&gt;I've gotten a bunch of questions from support after opening the case, should I continue to answer all of those right away or should I wait for your confirmation of whether this is indeed the issue you're mentioning?&lt;/P&gt;</description>
      <pubDate>Thu, 18 Aug 2022 16:40:43 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610009#M105728</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-08-18T16:40:43Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610013#M105729</link>
      <description>&lt;P&gt;I will answer depending on how the diag looks. But the information we have seen so far is "Shutdown control" signal received by HF is indicating known issue. But it's always good to see complete raw splunkd.log for confirmation.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 18 Aug 2022 17:49:04 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610013#M105729</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-08-18T17:49:04Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610117#M105732</link>
      <description>&lt;P&gt;Thanks.&lt;/P&gt;&lt;P&gt;Quick question: Is this a new bug in 9.x, or was this present in 8.x also?&lt;BR /&gt;I want to hold off upgrading the rest of our Environment (indexers etc.) if this bug was introduced in 9.x, we're still running 8.x on most Splunk components.&lt;/P&gt;</description>
      <pubDate>Fri, 19 Aug 2022 11:52:41 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/610117#M105732</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-08-19T11:52:41Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/622002#M107091</link>
      <description>&lt;P&gt;Ok we found a known issue that is only applicable for following workflow.&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;gt;Data flow is as follows:&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;&amp;gt;UFs sending logs with [httpout] -&amp;gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;New httpout when local UF is shutting down is able to pass down shutdown signal to next layer(HF/IDX). Causing to shutdown the pipeline that received it.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;It been fixed by&amp;nbsp;SPL-210081. Back ported to&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;8.1.8/&lt;/SPAN&gt;&lt;SPAN&gt;8.2.5.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 24 Nov 2022 14:04:54 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/622002#M107091</guid>
      <dc:creator>hrawat</dc:creator>
      <dc:date>2022-11-24T14:04:54Z</dc:date>
    </item>
    <item>
      <title>Re: Why do our Heavy Forwarders randomly shut down one of its parallel pipelines?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/622053#M107098</link>
      <description>&lt;P&gt;Thanks for the feedback!&lt;BR /&gt;&lt;BR /&gt;It's a bit strange though, since we were told by support at&amp;nbsp;November 4th in our support case that it was fixed by SPL-224974, which is a fix for a vulnerability listed here:&amp;nbsp;&lt;A href="https://www.splunk.com/en_us/product-security/announcements/svd-2022-1112.html" target="_blank"&gt;https://www.splunk.com/en_us/product-security/announcements/svd-2022-1112.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;We were also told that the versions that includes the fix is 9.0.2, 8.1.12 and 8.2.9.&lt;BR /&gt;&lt;SPAN&gt;&lt;BR /&gt;Maybe it is two different bugs?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Either way, I haven't been able to test this with any of the new versions yet, and I probably wont until after new year.&lt;/P&gt;</description>
      <pubDate>Fri, 25 Nov 2022 06:46:40 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Why-do-our-Heavy-Forwarders-randomly-shut-down-one-of-its/m-p/622053#M107098</guid>
      <dc:creator>mortenklow</dc:creator>
      <dc:date>2022-11-25T06:46:40Z</dc:date>
    </item>
  </channel>
</rss>

