<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Internal Index volume in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522586#M88267</link>
    <description>Hardly said any exact number as there are lot of things which need to count in. E.g. how many UF, HF, other inputs, which kind of distributed environment you have, how many and how actively they are using it etc.&lt;BR /&gt;You should look what is normal for your own environment with MC or use some queries for that.&lt;BR /&gt;r. Ismo</description>
    <pubDate>Thu, 01 Oct 2020 20:01:40 GMT</pubDate>
    <dc:creator>isoutamo</dc:creator>
    <dc:date>2020-10-01T20:01:40Z</dc:date>
    <item>
      <title>Internal Index volume</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522583#M88266</link>
      <description>&lt;P&gt;Just for a sake of knowledge, how much amount of _internal data is generated.&lt;/P&gt;&lt;P&gt;Incase my daily indexing is of 6TB???&lt;/P&gt;&lt;P&gt;Will it 15% of 6TB?&lt;/P&gt;&lt;P&gt;I know it doesn't consume my license...&lt;/P&gt;</description>
      <pubDate>Thu, 01 Oct 2020 19:42:36 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522583#M88266</guid>
      <dc:creator>hectorvp</dc:creator>
      <dc:date>2020-10-01T19:42:36Z</dc:date>
    </item>
    <item>
      <title>Re: Internal Index volume</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522586#M88267</link>
      <description>Hardly said any exact number as there are lot of things which need to count in. E.g. how many UF, HF, other inputs, which kind of distributed environment you have, how many and how actively they are using it etc.&lt;BR /&gt;You should look what is normal for your own environment with MC or use some queries for that.&lt;BR /&gt;r. Ismo</description>
      <pubDate>Thu, 01 Oct 2020 20:01:40 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522586#M88267</guid>
      <dc:creator>isoutamo</dc:creator>
      <dc:date>2020-10-01T20:01:40Z</dc:date>
    </item>
    <item>
      <title>Re: Internal Index volume</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522610#M88269</link>
      <description>&lt;P&gt;Hi &lt;a href="https://community.splunk.com/t5/user/viewprofilepage/user-id/214410"&gt;@isoutamo&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;We are currently having 500UFs and no HFs and scope is to fetch only UFs internal logs.&lt;/P&gt;&lt;P&gt;We are using single box for indexing,search head and as a DS, coz main purpose is to forward logs to 3rd party destn servers.&lt;/P&gt;&lt;P&gt;So we are suppose to only store internal logs, need to plan how much disk space will be required, currently we opting with 1TB but I guess this capacity planning has to be revisited. We need atleast retention policy of 60days.&lt;/P&gt;&lt;P&gt;Any suggestions on this, what could be the approach, we don't have actual access to 3rd party servers to check volume by using queries.&lt;/P&gt;</description>
      <pubDate>Fri, 02 Oct 2020 06:22:59 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522610#M88269</guid>
      <dc:creator>hectorvp</dc:creator>
      <dc:date>2020-10-02T06:22:59Z</dc:date>
    </item>
    <item>
      <title>Re: Internal Index volume</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522839#M88305</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.splunk.com/t5/user/viewprofilepage/user-id/214410"&gt;@isoutamo&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;I did the following to make rough estimate that is&lt;/P&gt;&lt;P&gt;avgerae size of internal logs = 300 bytes&amp;nbsp;&lt;/P&gt;&lt;P&gt;average events per seconds by internal index with 1 UF = 10&lt;/P&gt;&lt;P&gt;total internal data at&amp;nbsp; 1 day in MB by 1UF= 10 * 300 * 60 *60 *24 / (1000*1000) = 260 MB&lt;/P&gt;&lt;P&gt;So for&amp;nbsp; &amp;nbsp;1 day with 500UFs = 26 * 500 = 130000= 130GB in&amp;nbsp; day&lt;/P&gt;&lt;P&gt;Compression ratio of 50% so total 1 day data =&amp;nbsp; 65GB&lt;/P&gt;&lt;P&gt;so 60 days of retention capacity then = 65*60 = approx 4TB&lt;/P&gt;&lt;P&gt;Am I going in right direction???&amp;nbsp; Or missing any factors?&lt;/P&gt;</description>
      <pubDate>Sun, 04 Oct 2020 11:54:48 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/Internal-Index-volume/m-p/522839#M88305</guid>
      <dc:creator>hectorvp</dc:creator>
      <dc:date>2020-10-04T11:54:48Z</dc:date>
    </item>
  </channel>
</rss>

