<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Indexing and Searching Performance issues in Splunk Enterprise Security</title>
    <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281667#M2324</link>
    <description>&lt;P&gt;If you're spending tons of time on the &lt;CODE&gt;indexer&lt;/CODE&gt; processor, I'd blindly blame the actual writing-to-disk being slow... assuming that's the indexers' processor, not the HFs.&lt;BR /&gt;
"Down there" in the pipeline, all the complicated stuff has already been done, all that's left is to write it to disk: &lt;A href="http://wiki.splunk.com/Community:HowIndexingWorks"&gt;http://wiki.splunk.com/Community:HowIndexingWorks&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;Does that affect all seven indexers at the same time, or is it "two overloaded, five idle"? If the former, blame storage again. If the latter, add more HFs to balance things out.&lt;/P&gt;

&lt;P&gt;Note, imbalanced indexing also can affect search performance. If indexers 1 and 2 get all the data to index, indexers 1 and 2 also have to serve all search requests on their own.&lt;/P&gt;</description>
    <pubDate>Tue, 12 Apr 2016 16:27:54 GMT</pubDate>
    <dc:creator>martin_mueller</dc:creator>
    <dc:date>2016-04-12T16:27:54Z</dc:date>
    <item>
      <title>Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281658#M2315</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;

&lt;P&gt;I'm writing here out of desperation. We're having significant performance issues with our Splunk environment. I'll share as much info as I can and welcome any input or suggestions greatly:&lt;/P&gt;

&lt;P&gt;2 standalone search heads&lt;BR /&gt;
 - 1 ES &lt;BR /&gt;
 - 1 non-ES Searching and Reporting&lt;/P&gt;

&lt;P&gt;7 indexers&lt;BR /&gt;
2 heavy forwarders&lt;/P&gt;

&lt;P&gt;~8000 UFs&lt;/P&gt;

&lt;P&gt;All boxes are 20 cores and 48 GB RAM running Ubuntu and on ESX in a dedicated UCS farm with no overprovisioning. We're using shared Vmax storage for indexers and shared NIMBLE everywhere else. &lt;/P&gt;

&lt;P&gt;All of our indexing and forwarding queues are 90+% filled and our indexing is hours and in some cases days behind. &lt;/P&gt;

&lt;P&gt;We're struggling to identify the root cause. Any feedback is hugely appreciated. &lt;/P&gt;

&lt;P&gt;Thank you in advance. &lt;/P&gt;</description>
      <pubDate>Mon, 11 Apr 2016 23:22:19 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281658#M2315</guid>
      <dc:creator>cbauerlein</dc:creator>
      <dc:date>2016-04-11T23:22:19Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281659#M2316</link>
      <description>&lt;P&gt;How much volume are you pushing to your 7 indexers? In ES environments, 100gb/indexer can be too much.&lt;/P&gt;

&lt;P&gt;Are you routing all 8000 UFs through two HFs? If so, only two indexers receive data at a time making bad use of the available boxes. Consider adding more HFs, or upgrade to 6.3+ and use multiple parallel output pipeline sets to send to multiple indexers per HF.&lt;/P&gt;

&lt;P&gt;What processors are the biggest in the indexing performance view of the distributed management console? For example, if regexreplacement takes lots of time you probably have inefficient regular expressions running at index time.&lt;BR /&gt;
Make sure to also include the HF's logs in this as those may be doing most of the heavy lifting during parsing.&lt;BR /&gt;
To further narrow this down, check where the full queues "end" - the last processors in the pipeline that have full input queues usually are the culprits.&lt;/P&gt;

&lt;P&gt;Is it just indexing, or also searching that is affected? If it's just indexing and you have plenty of idle cores, you can consider upgrading to 6.3+ and use multiple parallel indexing pipeline sets to speed up indexing at the expense of using more cores.&lt;/P&gt;

&lt;P&gt;In general, update Splunk and especially ES and any standard TAs you have to current versions. Stuff like windows, unix, oracle, etc. TAs have recently been updated with greatly improved performance. Splunk 6.4.0 has also made some improvements to core itself for handling ES type workloads more efficiently.&lt;/P&gt;</description>
      <pubDate>Mon, 11 Apr 2016 23:36:08 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281659#M2316</guid>
      <dc:creator>martin_mueller</dc:creator>
      <dc:date>2016-04-11T23:36:08Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281660#M2317</link>
      <description>&lt;P&gt;If using ES I would target more to 70-80GB/day per indexer. On 6.3+ I would definitely add more HFs with an input pipeline of 2 in front of your indexers. Then do the same pipeline improvement on the indexers once you get more HFs added. You might also consider forcing time based load balancing from the HFs to the Indexers.&lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 00:06:51 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281660#M2317</guid>
      <dc:creator>starcher</dc:creator>
      <dc:date>2016-04-12T00:06:51Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281661#M2318</link>
      <description>&lt;P&gt;All systems are virtual?  You mentioned no over-provisioning, but do you actually have reserved CPU/Memory allocated in ESX?  Make sure the reservations are explicitly set per guest system.&lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 01:25:29 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281661#M2318</guid>
      <dc:creator>joshuascott94</dc:creator>
      <dc:date>2016-04-12T01:25:29Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281662#M2319</link>
      <description>&lt;P&gt;Hi - we do about 350 GB per day. &lt;/P&gt;

&lt;P&gt;Yes all UFs are sent directly to HFs. We are running 6.3.1 and ES 4. &lt;/P&gt;

&lt;P&gt;I will have to get back to you on the processors and queues.&lt;/P&gt;

&lt;P&gt;Both searching and indexing are issues. My main concern is indexing though because I don't want to be in a situation where I'm losing data. &lt;/P&gt;

&lt;P&gt;Searching is extremely slow and 90% of CPU load on the indexers is reduced when powering down the SHs. &lt;/P&gt;

&lt;P&gt;We've been told by colleagues that our shared storage could potentially be the culprit here but I just find that hard to believe. &lt;/P&gt;

&lt;P&gt;Thanks for your quick responses. &lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 01:45:52 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281662#M2319</guid>
      <dc:creator>cbauerlein</dc:creator>
      <dc:date>2016-04-12T01:45:52Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281663#M2320</link>
      <description>&lt;P&gt;Regarding storage being the culprit:  what type of CPU load are you seeing when looking at your process list in 'top' or 'vmstat'?  Is it mostly USER time or is it SYSTEM?  If you are seeing a lot of CPU time spent on the SYSTEM side, you could be IO bound.  If it is all tied up in USER land you are CPU bound.  For IO-bound workloads, you will need to increase your available storage IOPS/bandwidth.  For CPU-bound workloads, add processors/cores.&lt;/P&gt;

&lt;P&gt;How many of your ES datamodels are you really using and how many do you have accelerated?  Try disabling acceleration on datamodels that you are not using.  You might also want to look at limiting the number of concurrent datamodel acceleration jobs that can run simultaneously as well as the backfill_time in your datamodels.conf:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;[default]
acceleration.max_concurrent = 1
acceleration.backfill_time = -1h
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Tue, 12 Apr 2016 13:02:46 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281663#M2320</guid>
      <dc:creator>vasildavid</dc:creator>
      <dc:date>2016-04-12T13:02:46Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281664#M2321</link>
      <description>&lt;P&gt;My indexers spend the most process time on indexer&lt;/P&gt;

&lt;P&gt;All of my data queues are 100% except parsing queues which range from 85 - 97% full.&lt;/P&gt;

&lt;P&gt;We have not set explicit reservations.  My VM team has assured us this is not necessary as we have one virtual host per blade and have left 4 cores and 16 GB mem for ESX.&lt;/P&gt;

&lt;P&gt;Here is the output from top on one of my indexers:&lt;/P&gt;

&lt;P&gt;PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND&lt;BR /&gt;
 2541 root      20   0 2037788 882972   8768 S 103.0  1.8  11:33.48 splunkd&lt;BR /&gt;
 8307 root      20   0  253304 110608   8156 S 100.3  0.2   3:59.04 splunkd&lt;BR /&gt;
20956 root      20   0 2100908 895840   8760 S 100.3  1.8 131:29.97 splunkd&lt;BR /&gt;
30105 root      20   0 2693144 1.257g   8744 S 100.3  2.7  42:34.67 splunkd&lt;BR /&gt;
 2185 root      20   0 1687216 385472   8992 S 100.0  0.8  11:22.53 splunkd&lt;BR /&gt;
23300 root      20   0  622288 532196   7500 S 100.0  1.1   0:53.69 splunkd&lt;BR /&gt;
23885 root      20   0 1175232 252336   7736 S 100.0  0.5   0:32.61 splunkd&lt;BR /&gt;
 8336 root      20   0  183676 110536   7304 S  99.7  0.2   3:59.64 splunkd&lt;BR /&gt;
23878 root      20   0 1350480 198208   7896 S  98.7  0.4   0:31.95 splunkd&lt;BR /&gt;
18148 root      20   0 1652732 159784   8172 S  98.0  0.3   2:14.20 splunkd&lt;BR /&gt;
26549 root      20   0 1445884 260176   8236 S  94.7  0.5  13:12.25 splunkd&lt;BR /&gt;
25556 root      20   0 2524964 450532  14332 S  59.1  0.9 578:43.78 splunkd&lt;BR /&gt;
24804 root      20   0  122236  47516   5488 S  21.6  0.1   0:00.65 splunkd&lt;BR /&gt;
25691 root      20   0   93304  11660   9112 S   0.7  0.0   1:29.14 splunkd&lt;BR /&gt;
    8 root      20   0       0      0      0 S   0.3  0.0  73:30.09 rcu_sched&lt;BR /&gt;
   15 root      20   0       0      0      0 S   0.3  0.0  10:30.61 rcuos/6&lt;BR /&gt;
   21 root      20   0       0      0      0 S   0.3  0.0  12:38.55 rcuos/12&lt;BR /&gt;
   22 root      20   0       0      0      0 S   0.3  0.0  11:49.96 rcuos/13&lt;BR /&gt;
   24 root      20   0       0      0      0 S   0.3  0.0  10:48.43 rcuos/15&lt;BR /&gt;
18158 root      20   0   62820   5856    528 S   0.3  0.0   0:00.39 splunkd&lt;BR /&gt;
21499 root      20   0   17100   1736   1084 R   0.3  0.0   0:00.23 top&lt;BR /&gt;
23879 root      20   0   62820   5868    528 S   0.3  0.0   0:00.08 splunkd&lt;/P&gt;

&lt;P&gt;and vmstat:&lt;/P&gt;

&lt;P&gt;procs -----------memory---------- ---swap-- -----io---- -system-- ------cpu-----&lt;BR /&gt;
 r  b   swpd   free   buff  cache   si   so    bi    bo   in   cs us sy id wa st&lt;BR /&gt;
16  1 144848 3053864     20 32453728    0    0    31   458    1    1 22  2 76  0  0&lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 13:41:58 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281664#M2321</guid>
      <dc:creator>cbauerlein</dc:creator>
      <dc:date>2016-04-12T13:41:58Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281665#M2322</link>
      <description>&lt;P&gt;Do you have transparent huge pages disabled on the systems?&lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 14:08:08 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281665#M2322</guid>
      <dc:creator>starcher</dc:creator>
      <dc:date>2016-04-12T14:08:08Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281666#M2323</link>
      <description>&lt;P&gt;Yes THP has been disabled and ulimit increased or set to unlimited.&lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 14:19:45 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281666#M2323</guid>
      <dc:creator>cbauerlein</dc:creator>
      <dc:date>2016-04-12T14:19:45Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281667#M2324</link>
      <description>&lt;P&gt;If you're spending tons of time on the &lt;CODE&gt;indexer&lt;/CODE&gt; processor, I'd blindly blame the actual writing-to-disk being slow... assuming that's the indexers' processor, not the HFs.&lt;BR /&gt;
"Down there" in the pipeline, all the complicated stuff has already been done, all that's left is to write it to disk: &lt;A href="http://wiki.splunk.com/Community:HowIndexingWorks"&gt;http://wiki.splunk.com/Community:HowIndexingWorks&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;Does that affect all seven indexers at the same time, or is it "two overloaded, five idle"? If the former, blame storage again. If the latter, add more HFs to balance things out.&lt;/P&gt;

&lt;P&gt;Note, imbalanced indexing also can affect search performance. If indexers 1 and 2 get all the data to index, indexers 1 and 2 also have to serve all search requests on their own.&lt;/P&gt;</description>
      <pubDate>Tue, 12 Apr 2016 16:27:54 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281667#M2324</guid>
      <dc:creator>martin_mueller</dc:creator>
      <dc:date>2016-04-12T16:27:54Z</dc:date>
    </item>
    <item>
      <title>Re: Indexing and Searching Performance issues</title>
      <link>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281668#M2325</link>
      <description>&lt;P&gt;Did you get this resolved? Have you limited your datamodels to only creating their summaries from the indexes they require? This saved us a huge amount of resource when we implemented it.&lt;/P&gt;</description>
      <pubDate>Tue, 11 Oct 2016 10:51:20 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Enterprise-Security/Indexing-and-Searching-Performance-issues/m-p/281668#M2325</guid>
      <dc:creator>kerryc</dc:creator>
      <dc:date>2016-10-11T10:51:20Z</dc:date>
    </item>
  </channel>
</rss>

