<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Splunk daemon stops responding when many connections are made in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219100#M43037</link>
    <description>&lt;P&gt;Update: All these errors occur after splunk daemon stops responding to requests. Upon further debugging, we found that when splunk daemon stops, the &lt;CODE&gt;rest.simpleRequest&lt;/CODE&gt; method which is called for authentication times out without closing the connection which results in many threads in CLOSE_WAIT state. We really want to know why exactly splunk daemon stops responding even when the &lt;CODE&gt;splunk status&lt;/CODE&gt; command shows that splunkd is running.&lt;/P&gt;</description>
    <pubDate>Tue, 15 Nov 2016 07:15:55 GMT</pubDate>
    <dc:creator>keerthana_k</dc:creator>
    <dc:date>2016-11-15T07:15:55Z</dc:date>
    <item>
      <title>Splunk daemon stops responding when many connections are made</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219098#M43035</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;

&lt;P&gt;We have a Splunk deployment where we expose a splunk app through REST. We connect to the app and run approximately 50 searches every 5 minutes. However, after a while we observe that the searches stop running. Upon debugging we found the following:&lt;/P&gt;

&lt;P&gt;In the webservice.log file, we got the following error:&lt;/P&gt;

&lt;P&gt;2016-11-06 06:28:32,809 ERROR   [581ecd72c57f75d01abbd0] &lt;STRONG&gt;init&lt;/STRONG&gt;:479 - Socket error communicating with splunkd (error=('&lt;EM&gt;ssl.c:587: The handshake operation timed out',)), path = &lt;A href="https://localhost:8089/services/auth/login" target="_blank"&gt;https://localhost:8089/services/auth/login&lt;/A&gt;&lt;BR /&gt;
2016-11-06 06:29:02,853 ERROR   [581ecd90d17f75d014d690] __init&lt;/EM&gt;_:479 - Socket error communicating with splunkd (error=('_ssl.c:587: The handshake operation timed out',)), path = &lt;A href="https://localhost:8089/services/auth/login" target="_blank"&gt;https://localhost:8089/services/auth/login&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;In the python.log file, we got:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;Splunkd daemon is not responding: ("Error connecting to https://localhost:8089/services/auth/login: ('_ssl.c:587: The handshake operation timed out',)",)
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;In splunkd.log file, we got the following error:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;11-13-2016 23:18:58.662 +0000 ERROR HttpListener - Exception while processing request from 1.1.1.1 for /en-US/custom/XXX/YYY/URLparameters: Connection closed by peer
11-13-2016 23:18:58.662 +0000 ERROR HttpListener - Handler for  /en-US/custom/XXX/YYY/URLparameters sent a 0 byte response after earlier claiming a Content-Length of 26!
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;We ran the command &lt;CODE&gt;netstat -neap | grep 8089&lt;/CODE&gt; and got the following output:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;tcp        0      0 0.0.0.0:8089                0.0.0.0:*                   LISTEN      0          136045514
tcp        0      0 127.0.0.1:8089              127.0.0.1:52914             SYN_RECV    0          0
tcp        0      0 127.0.0.1:8089              127.0.0.1:52908             SYN_RECV    0          0
tcp        1      0 127.0.0.1:8089              127.0.0.1:45149             CLOSE_WAIT  0          139900826     14993/splunkd 
tcp        1      0 127.0.0.1:8089              127.0.0.1:45984             CLOSE_WAIT  0          140090123     14993/splunkd 
tcp        1      0 127.0.0.1:8089              127.0.0.1:42823             CLOSE_WAIT  0          140067803     14993/splunkd 
tcp        1      0 127.0.0.1:8089              127.0.0.1:33117             CLOSE_WAIT  0          140194381     14993/splunkd 
tcp        1      0 127.0.0.1:8089              127.0.0.1:48657             CLOSE_WAIT  0          139922911     14993/splunkd 
tcp        1      0 127.0.0.1:8089              127.0.0.1:60258             CLOSE_WAIT  0          140186060     14993/splunkd 
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;The ulimit value for the system is set as 4096 and we found event in splunkd.log saying:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;11-14-2016 04:13:25.605 +0000 INFO  loader - Limiting REST HTTP server to 1365 sockets
11-14-2016 04:13:25.605 +0000 INFO  loader - Limiting REST HTTP server to 1365 threads 
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;How do we fix this issue? When does the CLOSE_WAIT status appear for any connection? Any help is greatly appreciated.&lt;/P&gt;

&lt;P&gt;Thanks,&lt;BR /&gt;
Keerthana&lt;/P&gt;</description>
      <pubDate>Tue, 29 Sep 2020 11:45:02 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219098#M43035</guid>
      <dc:creator>keerthana_k</dc:creator>
      <dc:date>2020-09-29T11:45:02Z</dc:date>
    </item>
    <item>
      <title>Re: Splunk daemon stops responding when many connections are made</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219099#M43036</link>
      <description>&lt;P&gt;CLOSE_WAIT is a TCP state, you can find an explanation online of what exactly that means.&lt;/P&gt;

&lt;P&gt;I would increase your ulimit for the number of file descriptors.&lt;/P&gt;

&lt;P&gt;&lt;A href="https://docs.splunk.com/Documentation/Splunk/6.5.0/Troubleshooting/ulimitErrors"&gt;Splunk ulimit troubleshooting documentation&lt;/A&gt; says 8192 minimum, I run much, much larger numbers on the indexers.&lt;BR /&gt;
4192 is not enough for Splunk...&lt;/P&gt;</description>
      <pubDate>Tue, 15 Nov 2016 06:42:18 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219099#M43036</guid>
      <dc:creator>gjanders</dc:creator>
      <dc:date>2016-11-15T06:42:18Z</dc:date>
    </item>
    <item>
      <title>Re: Splunk daemon stops responding when many connections are made</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219100#M43037</link>
      <description>&lt;P&gt;Update: All these errors occur after splunk daemon stops responding to requests. Upon further debugging, we found that when splunk daemon stops, the &lt;CODE&gt;rest.simpleRequest&lt;/CODE&gt; method which is called for authentication times out without closing the connection which results in many threads in CLOSE_WAIT state. We really want to know why exactly splunk daemon stops responding even when the &lt;CODE&gt;splunk status&lt;/CODE&gt; command shows that splunkd is running.&lt;/P&gt;</description>
      <pubDate>Tue, 15 Nov 2016 07:15:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219100#M43037</guid>
      <dc:creator>keerthana_k</dc:creator>
      <dc:date>2016-11-15T07:15:55Z</dc:date>
    </item>
    <item>
      <title>Re: Splunk daemon stops responding when many connections are made</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219101#M43038</link>
      <description>&lt;P&gt;This issue could still occur after we increase the ulimit when the number of file descriptors reaches 33% of set ulimit value. &lt;/P&gt;</description>
      <pubDate>Tue, 15 Nov 2016 07:18:09 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219101#M43038</guid>
      <dc:creator>keerthana_k</dc:creator>
      <dc:date>2016-11-15T07:18:09Z</dc:date>
    </item>
    <item>
      <title>Re: Splunk daemon stops responding when many connections are made</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219102#M43039</link>
      <description>&lt;P&gt;The issue would be due to Global Timeout parameters set in some python libraries.&lt;/P&gt;

&lt;P&gt;Its better to set timeout explicitly.&lt;/P&gt;

&lt;P&gt;Refer: &lt;A href="https://answers.splunk.com/answers/477671/how-to-resolve-lots-of-splunk-internal-connections.html"&gt;https://answers.splunk.com/answers/477671/how-to-resolve-lots-of-splunk-internal-connections.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 06 Jan 2017 10:11:11 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Splunk-daemon-stops-responding-when-many-connections-are-made/m-p/219102#M43039</guid>
      <dc:creator>strive</dc:creator>
      <dc:date>2017-01-06T10:11:11Z</dc:date>
    </item>
  </channel>
</rss>

