<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: search heads failing because of huge knowledge bundles in Deployment Architecture</title>
    <link>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417434#M20456</link>
    <description>&lt;P&gt;when i ran the commands as suggested by you, i got the below results, i was of the view that &lt;BR /&gt;
2.6G    ./apps&lt;BR /&gt;
2.6G    .&lt;BR /&gt;
328K    ./system&lt;BR /&gt;
56K     ./users&lt;BR /&gt;
48K     ./kvstore_s_SA-&lt;BR /&gt;
is it safe to blacklist the apps directory entirely, we have a huge dependency on the TA and app for AWS. on further troubleshooting i found that the lookup aws_description.csv is taking up close to 2.3 GB. is it safe to blacklist the aws_description.csv lookup, since we would require aws description data for alerts and reports.&lt;BR /&gt;
In case i need to blacklist, will the below setting work&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;[replicationBlacklist]
&amp;lt;name for lookup directories&amp;gt; = (.../lookups/...)
&amp;lt;name for bin and jardirectories&amp;gt; = (.../(bin|jars)/...)
&lt;/CODE&gt;&lt;/PRE&gt;</description>
    <pubDate>Tue, 29 Sep 2020 20:58:29 GMT</pubDate>
    <dc:creator>soumyasaha25</dc:creator>
    <dc:date>2020-09-29T20:58:29Z</dc:date>
    <item>
      <title>search heads failing because of huge knowledge bundles</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417432#M20454</link>
      <description>&lt;P&gt;currently half of my searchheads are shutdown (auto shutdown due to issues within Splunk) and the remaining are not able to query the indexers&lt;BR /&gt;
The problem is caused by a large knowledge bundle. &lt;BR /&gt;
when i checked the .bundle files on the SHs, it is a huge (~340 MB) file with what looks like a huge python code.&lt;BR /&gt;
i have maxBundleSize set to 2048(which is the default)&lt;BR /&gt;
i have a blacklist in distsearch.conf which is as below:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;[replicationSettings]
maxBundleSize = 2048

[replicationBlacklist]
&amp;lt;name for bin directories&amp;gt; = (.../bin/*)
&amp;lt;name for InstallDirectories&amp;gt; = (.../install/*)
&amp;lt;name for AppServerDirectories&amp;gt; = (.../appserver/*)
&amp;lt;name for allAppUIDirectories&amp;gt; = (.../default/data/ui/*)
&amp;lt;name for allOldDefaultDirectories&amp;gt; = (.../default.old.*)
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;My questions is: is there any way to check what files/apps are included in this bundle that is causing issues and if those items are required or can be excluded.&lt;/P&gt;</description>
      <pubDate>Tue, 21 Aug 2018 15:51:40 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417432#M20454</guid>
      <dc:creator>soumyasaha25</dc:creator>
      <dc:date>2018-08-21T15:51:40Z</dc:date>
    </item>
    <item>
      <title>Re: search heads failing because of huge knowledge bundles</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417433#M20455</link>
      <description>&lt;P&gt;mkdir -p /tmp/support&lt;BR /&gt;
tar xvf /opt/splunk/var/run/blah.bundle -C /tmp/support&lt;BR /&gt;
cd /tmp/support&lt;BR /&gt;
du -h --max-depth=1 |sort -hr |more&lt;/P&gt;

&lt;P&gt;Walk it out most likley be in apps/*/lookups&lt;/P&gt;

&lt;P&gt;You can blacklist any lookup that is not:&lt;/P&gt;

&lt;OL&gt;
&lt;LI&gt;Automatic Lookup "props.conf"&lt;/LI&gt;
&lt;LI&gt;any csv that is not being searched with a |lookup .  because that is a remote lookup...If you are using | lookup local=true you could blacklist it.&lt;/LI&gt;
&lt;/OL&gt;

&lt;P&gt;And if you blacklist it..... and you get an error after the fact.  Unblacklist it.&lt;/P&gt;

&lt;P&gt;Don't forget that you have a Transmit side and a Receive side&lt;/P&gt;

&lt;P&gt;In your case the transmit side is your SH and the distsearch.conf setting maxBundleSize applies&lt;BR /&gt;
however&lt;BR /&gt;
the receive side is your indexers... and that setting is server.conf&lt;BR /&gt;
[httpServer]&lt;BR /&gt;
max_content_length = blah&lt;/P&gt;

&lt;P&gt;And depending on your version it might be 800mb or 2gb but written as 2147483648 in the 2gb example.&lt;/P&gt;

&lt;P&gt;Hope this helps.&lt;/P&gt;</description>
      <pubDate>Tue, 29 Sep 2020 20:58:10 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417433#M20455</guid>
      <dc:creator>jwelch_splunk</dc:creator>
      <dc:date>2020-09-29T20:58:10Z</dc:date>
    </item>
    <item>
      <title>Re: search heads failing because of huge knowledge bundles</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417434#M20456</link>
      <description>&lt;P&gt;when i ran the commands as suggested by you, i got the below results, i was of the view that &lt;BR /&gt;
2.6G    ./apps&lt;BR /&gt;
2.6G    .&lt;BR /&gt;
328K    ./system&lt;BR /&gt;
56K     ./users&lt;BR /&gt;
48K     ./kvstore_s_SA-&lt;BR /&gt;
is it safe to blacklist the apps directory entirely, we have a huge dependency on the TA and app for AWS. on further troubleshooting i found that the lookup aws_description.csv is taking up close to 2.3 GB. is it safe to blacklist the aws_description.csv lookup, since we would require aws description data for alerts and reports.&lt;BR /&gt;
In case i need to blacklist, will the below setting work&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;[replicationBlacklist]
&amp;lt;name for lookup directories&amp;gt; = (.../lookups/...)
&amp;lt;name for bin and jardirectories&amp;gt; = (.../(bin|jars)/...)
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Tue, 29 Sep 2020 20:58:29 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417434#M20456</guid>
      <dc:creator>soumyasaha25</dc:creator>
      <dc:date>2020-09-29T20:58:29Z</dc:date>
    </item>
    <item>
      <title>Re: search heads failing because of huge knowledge bundles</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417435#M20457</link>
      <description>&lt;P&gt;I deleted my last post because I missed your part about the aws_description.csv being 2.3 GB.&lt;/P&gt;

&lt;P&gt;As I mentioned earlier.....   You need to find out if that file is being used as part of an automatic lookup in a props statement.  If it is not blacklist the file.  If you get errors after the fact un-blacklist it.&lt;/P&gt;

&lt;P&gt;And figure out why that csv is so big.  You might want to file a support case and work with an AWS SME.&lt;/P&gt;

&lt;P&gt;Bottom line is if the lookup is being performed on the SH you don't need the CSV in the bundle.&lt;/P&gt;

&lt;P&gt;If you find you do need it, then you need to increase your maxBundleSize and max_content_length, but I would suspect something is wrong if that file is 2.3 gb&lt;/P&gt;</description>
      <pubDate>Tue, 29 Sep 2020 20:58:42 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417435#M20457</guid>
      <dc:creator>jwelch_splunk</dc:creator>
      <dc:date>2020-09-29T20:58:42Z</dc:date>
    </item>
    <item>
      <title>Re: search heads failing because of huge knowledge bundles</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417436#M20458</link>
      <description>&lt;P&gt;Thanks a lot for your response.&lt;/P&gt;</description>
      <pubDate>Wed, 22 Aug 2018 14:24:11 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/search-heads-failing-because-of-huge-knowledge-bundles/m-p/417436#M20458</guid>
      <dc:creator>soumyasaha25</dc:creator>
      <dc:date>2018-08-22T14:24:11Z</dc:date>
    </item>
  </channel>
</rss>

