No errors in the monitoring console. I see this only in distributed setups, so no errors when searching locally on a all-in-one box. The worst installation has 1500+ broken pipes on these (search: 😞 09-09-2021 13:32:18.574 +0200 WARN HttpListener - Socket error from "search-head-ip":42800 while accessing /services/streams/search: Broken pipe and this type ERROR: 09-09-2021 13:40:33.742 +0200 ERROR HttpListener - Exception while processing request from "search-head-ip":45930 for /services/search/jobs/remote_splunksh03."domainname"_subsearch_nested_8294df4a2b86339a_1631187558.3/search.log: Broken pipe Other installations have between 20 and 50 of those pr 24 hours. These warnings and errors are logged a lot when searches start having errors like this (from search.log): 09-09-2021 14:00:45.634 ERROR SearchResultTransaction - Got status 502 from https://"indexer-ip":8089/services/streams/search?sh_sid=1631188834.20962_7D6DF087-C582-4B67-A82D-BD1F18B5BEA5
09-09-2021 14:00:45.634 INFO XmlParser - Entity: line 1: parser: Document is empty
09-09-2021 14:00:45.635 ERROR SearchResultParser - HTTP error status message from https://"indexer-ip":8089/services/streams/search?sh_sid=1631188834.20962_7D6DF087-C582-4B67-A82D-BD1F18B5BEA5: Error connecting: Connect Timeout
09-09-2021 14:00:45.635 WARN SearchResultParserExecutor - Error connecting: Connect Timeout for collector=splunkidx01.domainname
09-09-2021 14:00:45.635 ERROR DispatchThread - sid:1631188834.20962_7D6DF087-C582-4B67-A82D-BD1F18B5BEA5 Unknown error for indexer: splunkidx01.domainname. Search Results might be incomplete! If this occurs frequently, check on the peer. The peers are always up, the ulimit, thread limit and socket limit are all OK when viewing in the splunkd log when splunk is starting. (It is systemd managed on an ubuntu 20.02LTS, so ulimit set there, and is 65535) The servers (sh's and indexers) are 80 cores (40 hyperthreaded, dual socket), 386 GB ram, pure SSD. Seach head cluster with 3 nodes Indexer cluster with 4 nodes. We see no other network errors. No problems with 9887 replication data in the indexer cluster, no problems ingesting data.
... View more