<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Filesystem for Splunk in Knowledge Management</title>
    <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65208#M7243</link>
    <description>&lt;P&gt;(started as a comment, but couldn't fit in 500 chars... sorry)&lt;/P&gt;

&lt;P&gt;A very quick read suggests that the cachecade may not provide the desired boost to Splunk's workload.  Per Dell's documentation it is a read cache only.  It may not help substantially with the RAID-5 performance penalty for less than full-stripe writes.  Splunk deals well with cold buckets on RAID-5 (because cold buckets are read-only), but much less so with hot - because there is a lot of random writes to update both rawdata and index tsidx data files.&lt;/P&gt;

&lt;P&gt;To optimize for performance and capacity, my personal preference would be to put all 16 drives in the R720, and use 8 in a (4+4) RAID-10 for the hot buckets and 8 in a (7+P) RAID-5 for cold buckets / operating system.  Use Linux LVM to help sort out filesystems appropriately.  If you really want to use the cachecade, put it on the RAID-5 volume and let the RAID-10 be dealt with using operating system filesystem cache in RAM.&lt;/P&gt;</description>
    <pubDate>Thu, 20 Sep 2012 04:55:22 GMT</pubDate>
    <dc:creator>dwaddle</dc:creator>
    <dc:date>2012-09-20T04:55:22Z</dc:date>
    <item>
      <title>Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65202#M7237</link>
      <description>&lt;P&gt;so, under a rhel6 and latest splunk, and likely sitting on 16 10k spindles raid-5, is there a filesystem best suited for the job given the fact that both syslog-ng files and splunk indexes are on the same filesystem?&lt;/P&gt;</description>
      <pubDate>Wed, 19 Sep 2012 18:29:59 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65202#M7237</guid>
      <dc:creator>cvajs</dc:creator>
      <dc:date>2012-09-19T18:29:59Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65203#M7238</link>
      <description>&lt;P&gt;First and foremost - NOT RAID-5&lt;/P&gt;

&lt;P&gt;Raid 10 (1+0) would be much much better.  Normal RHEL filesystem s/b fine.&lt;/P&gt;

&lt;P&gt;Very good info in first part of Installation manual, and in wiki:&lt;/P&gt;

&lt;P&gt;&lt;A href="http://wiki.splunk.com/Community:HardwareTuningFactors"&gt;http://wiki.splunk.com/Community:HardwareTuningFactors&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;&lt;A href="http://docs.splunk.com/Documentation/Splunk/latest/Installation/Systemrequirements"&gt;http://docs.splunk.com/Documentation/Splunk/latest/Installation/Systemrequirements&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Sep 2012 21:23:56 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65203#M7238</guid>
      <dc:creator>lguinn2</dc:creator>
      <dc:date>2012-09-19T21:23:56Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65204#M7239</link>
      <description>&lt;P&gt;thanks for the info, but i am not 100% convinced that raid-5 is a bad choice. raid-5 is certainly not better than a flavor of raid-10, but if a 14 spindle raid-5 array offers ~20000+ IOPS does it really matter that it's a raid-5? the cons for flavor of raid-10 is that you lose half the storage capacity. pros and cons to each, but i am not ready to say raid-5 is a bad choice.&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 01:41:25 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65204#M7239</guid>
      <dc:creator>cvajs</dc:creator>
      <dc:date>2012-09-20T01:41:25Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65205#M7240</link>
      <description>&lt;P&gt;How can you get 20000+ IOPS from 10K spindles? Did you mean SSD?&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 01:46:08 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65205#M7240</guid>
      <dc:creator>reed_kelly</dc:creator>
      <dc:date>2012-09-20T01:46:08Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65206#M7241</link>
      <description>&lt;P&gt;in this specific case i am referencing Dell's R720 with one SSD CacheCade, tests ran in raid-10 using 15k drives yielded 32k+ IOPS. its a skewed # due to CacheCade and it all depends. 20k was my estimate based on same tests using 10k drives.&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 02:27:08 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65206#M7241</guid>
      <dc:creator>cvajs</dc:creator>
      <dc:date>2012-09-20T02:27:08Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65207#M7242</link>
      <description>&lt;P&gt;if i run crude calc using 1/3 read and 2/3 writes in raid-5 w/ 1k IOPS as my target for operating IOPS, i get something like this:&lt;BR /&gt;
(1000*0.333)+(1000*0.666*4)&lt;BR /&gt;
333+2664 = 2997 raw array IOPS to realize 1000 in raid-5&lt;/P&gt;

&lt;P&gt;14*130*cachecade factor = 1820*3.5 = 6370&lt;/P&gt;

&lt;P&gt;so, probably not 20k, but if the OS can "see" 6370 IOPS in a raid-5 split 33% read and 66% write, does my original statement still hold, does it really matter that its raid-5?&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 02:40:03 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65207#M7242</guid>
      <dc:creator>cvajs</dc:creator>
      <dc:date>2012-09-20T02:40:03Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65208#M7243</link>
      <description>&lt;P&gt;(started as a comment, but couldn't fit in 500 chars... sorry)&lt;/P&gt;

&lt;P&gt;A very quick read suggests that the cachecade may not provide the desired boost to Splunk's workload.  Per Dell's documentation it is a read cache only.  It may not help substantially with the RAID-5 performance penalty for less than full-stripe writes.  Splunk deals well with cold buckets on RAID-5 (because cold buckets are read-only), but much less so with hot - because there is a lot of random writes to update both rawdata and index tsidx data files.&lt;/P&gt;

&lt;P&gt;To optimize for performance and capacity, my personal preference would be to put all 16 drives in the R720, and use 8 in a (4+4) RAID-10 for the hot buckets and 8 in a (7+P) RAID-5 for cold buckets / operating system.  Use Linux LVM to help sort out filesystems appropriately.  If you really want to use the cachecade, put it on the RAID-5 volume and let the RAID-10 be dealt with using operating system filesystem cache in RAM.&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 04:55:22 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65208#M7243</guid>
      <dc:creator>dwaddle</dc:creator>
      <dc:date>2012-09-20T04:55:22Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65209#M7244</link>
      <description>&lt;P&gt;What dwaddle said!! Thanks dude &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 05:01:38 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65209#M7244</guid>
      <dc:creator>lguinn2</dc:creator>
      <dc:date>2012-09-20T05:01:38Z</dc:date>
    </item>
    <item>
      <title>Re: Filesystem for Splunk</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65210#M7245</link>
      <description>&lt;P&gt;so, 2.4T raid-10 for hot buckets, and 5+P+cachecade (need one spare) 3T raid-5 for OS and cold buckets. problem is, i have syslog on same system and i expect 60/40 write:read ratio between syslog writing and Splunk reading those files, so i suspect syslog has to go on raid-10. 2.4T wont be enough for my syslog data. i also plan to do some ext4 filesystem tuning, like noatime, data=ordered, nouser_xattr, which saves some writes.&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2012 14:59:13 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/Filesystem-for-Splunk/m-p/65210#M7245</guid>
      <dc:creator>cvajs</dc:creator>
      <dc:date>2012-09-20T14:59:13Z</dc:date>
    </item>
  </channel>
</rss>

