<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: SmartStore Behaviors in Knowledge Management</title>
    <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402776#M3500</link>
    <description>&lt;P&gt;The entirety of cold storage has 0 files in it. Remains true after running historical searches that would surely end up with some cold data in play.&lt;/P&gt;</description>
    <pubDate>Tue, 11 Jun 2019 20:28:03 GMT</pubDate>
    <dc:creator>twinspop</dc:creator>
    <dc:date>2019-06-11T20:28:03Z</dc:date>
    <item>
      <title>SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402765#M3489</link>
      <description>&lt;P&gt;I'd like to better understand what behaviors SmartStore is going to exhibit in my environment, and how do I manage them? What can I do to prepare my environment for SmartStore?&lt;/P&gt;</description>
      <pubDate>Mon, 08 Apr 2019 22:46:20 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402765#M3489</guid>
      <dc:creator>davidpaper</dc:creator>
      <dc:date>2019-04-08T22:46:20Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402766#M3490</link>
      <description>&lt;P&gt;S2 behaviors in no particular order. I will update this post as new information is learned.&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;RF/SF only apply to Hot buckets. Once a bucket is rolled, it is uploaded to S3 and any bucket replicates are marked for eviction.&lt;/LI&gt;&lt;LI&gt;S2 cachemanager will download components of a bucket as searches determine what’s needed. Maybe bloomfilters, deletes, journal.* or other components, and as such multiple downloads for the same bucket may look like they are happening, but per component, no duplicate downloads should happen.&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;Evictions don’t always seem to show up in MC on the S2 pages. The following will.&lt;/P&gt;&lt;P&gt;index=_internal sourcetype=splunkd source=*splunkd.log action=evictDeletes&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;Starting in 7.2.4, additional metrics were added to be able to count downloaded byte count. Prior to this version, Splunk was metrics-blind to the (potentially significant) impact on the network/storage a rolling restart induces.&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;During a rolling restart, as each indexer is marked to go down&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;CM begins to reassign primacy for buckets on the indexer on the way down to other indexers&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;All buckets on indexer being restarted are marked for eviction, effectively flushing the cache on the indexer being restarted&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;As indexers in the cluster are restarted, others will start d/ling buckets from S3 to satisfy search requests, which can take a heavy toll on local network and storage if not prepared for this level of data transfer in a short period of time, as all other indexers not being restarted will likely start requesting buckets to download at once.&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;SmartStore only allows one indexer at a time to be primary searchable for a bucket and no other indexers are allowed to have copies of that bucket cached. The CM will issue eviction notices to any indexers with copies of that bucket locally. This ensures that only 1 indexer will search that bucket and return results. As a result of this, there is a huge amount of data shuffling and downloading that happens during a full cluster rolling restart.&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;Bucket rebalance works more quickly with S2 than without it because the only buckets to rebalance are hot buckets&lt;/P&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;EM&gt;Added Nov 2019&lt;/EM&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Disk part 1: S2 disk I/O requirements seem to be higher than non-S2, due to the bucket downloading process needing to be able to write large amounts of data quickly as cachemanager populates buckets for search. Default downloading config allows for 8 simultaneous downloads at once. Disks previously able to shoulder the load may not be up to the task of S2’s caching requirements. &lt;STRONG&gt;I'm looking at you, RAID5 volumes.&lt;/STRONG&gt; By definition it's cache space (and hot bucket space, but hot is replicated), so use RAID0 (stripe) for the fastest disk possible, and not waste a MB of available disk space. RAID10 (mirrored stripes) is also acceptable, but cuts usable disk space by 50%.&lt;/LI&gt;&lt;LI&gt;Disk part 2: To expand on the above a bit, S2 performance is more than just high IOPS, it's about throughput too. Customers running S2 in AWS that have chosen to use gp2 EBS volumes for hot/cachemanager will likely see severe IO contention resulting in IO wait % jumping during high periods of S2 bucket downloads from remote storage. This is quite easy to see in top or iostat when users run searches that trigger large bucket evictions &amp;amp; bucket downloads from remote storage. gp2 has a limit of 250MB/sec, which doesn't take long to hit when the network is 10 gig or faster. Yes, a fast network means data written to kernel buffer cache at a high rate and when its time to sync to disk, the storage won't be able to keep up. io1 EBS type is better, at 1000MB/s, but still can exhaust throughput capacity during periods of concurrent high bucket downloads and search that taxes the storage for both reads and writes in addition to ingestion and hot bucket replication. In AWS, it is highly recommended to use NVME for hot/cachemanager (i3 and i3en instance types work very well here) in RAID 0 and consider setting RF/SF=3 (still applies to hot buckets) to sleep better at night.&lt;/LI&gt;&lt;LI&gt;Disk part 3: If deploying S2 outside of AWS, strive to obtain the fastest disks (throughput &amp;amp; IOPS) available, whether local SSDs or NVME to avoid storage bottlenecks getting in the way of your Splunk performance.&lt;/LI&gt;&lt;/UL&gt;</description>
      <pubDate>Thu, 13 May 2021 03:59:10 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402766#M3490</guid>
      <dc:creator>davidpaper</dc:creator>
      <dc:date>2021-05-13T03:59:10Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402767#M3491</link>
      <description>&lt;P&gt;This is a really good rundown for anyone planning to use S2. Thanks for the summary @davidpaper!&lt;/P&gt;</description>
      <pubDate>Mon, 08 Apr 2019 23:58:35 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402767#M3491</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2019-04-08T23:58:35Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402768#M3492</link>
      <description>&lt;P&gt;I think that it is worth noting explicitly (@davidpaper implied it) that currently, &lt;CODE&gt;SS/S3&lt;/CODE&gt; is currently &lt;STRONG&gt;NOT&lt;/STRONG&gt; practical for &lt;CODE&gt;hot/cold&lt;/CODE&gt; buckets/volumes and that it should &lt;STRONG&gt;ONLY&lt;/STRONG&gt; be used for &lt;CODE&gt;warm&lt;/CODE&gt;.&lt;/P&gt;</description>
      <pubDate>Sat, 13 Apr 2019 16:51:01 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402768#M3492</guid>
      <dc:creator>woodcock</dc:creator>
      <dc:date>2019-04-13T16:51:01Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402769#M3493</link>
      <description>&lt;P&gt;I'm not sure I follow. You don't have a choice of WARM or COLD with S2. There is HOT; briefly there is WARM while waiting to upload to remote; and finally there is remote with cached local copies. Th entire bucket lifecycle changes.&lt;/P&gt;

&lt;P&gt;At least this is my understanding.&lt;/P&gt;</description>
      <pubDate>Sun, 14 Apr 2019 01:06:39 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402769#M3493</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2019-04-14T01:06:39Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402770#M3494</link>
      <description>&lt;P&gt;Ah, this isn't really the case, but I can see how it might appear this way. There is now only "hot" and "not hot" in terms of a bucket lifecycle in S2. The concept of warm and cold being separate is no longer really a thing.&lt;/P&gt;

&lt;P&gt;Hot (read/write) is still replicated based on CM RF/SF settings until it rolls to read-only, and then 1 copy is made of the bucket to S3, and the other local copies are marked for deletion by the indexers' cachemanager process.&lt;/P&gt;

&lt;P&gt;The cachemanager retrieves read-only buckets from S3 when it needs to so a search can be completed and those bucket share the same file system as hot...so make sure your hot/cachemanager filesystem is nice and fast.&lt;/P&gt;</description>
      <pubDate>Wed, 05 Jun 2019 19:46:12 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402770#M3494</guid>
      <dc:creator>davidpaper</dc:creator>
      <dc:date>2019-06-05T19:46:12Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402771#M3495</link>
      <description>&lt;P&gt;After learning more, apparently a more proper statement is &lt;CODE&gt;When using SmartStore, there is no need to use cold at all and Splunk should always configured  to have NO COLD&lt;/CODE&gt;, or maybe not...?&lt;/P&gt;</description>
      <pubDate>Sun, 09 Jun 2019 21:51:58 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402771#M3495</guid>
      <dc:creator>woodcock</dc:creator>
      <dc:date>2019-06-09T21:51:58Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402772#M3496</link>
      <description>&lt;P&gt;Once an index is converted to use SmartStore, you are spot on. &lt;EM&gt;No more need for a coldPath entry for that index.&lt;/EM&gt;&lt;/P&gt;

&lt;P&gt;Edit: The above is incorrect. You still need a coldPath entry in indexes.conf for the index, but the cold volume shouldn't be actively used once the buckets have been evicted from there. &lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2019 13:59:23 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402772#M3496</guid>
      <dc:creator>davidpaper</dc:creator>
      <dc:date>2019-06-11T13:59:23Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402773#M3497</link>
      <description>&lt;P&gt;The index still requires a configured coldPath.  See &lt;A href="https://docs.splunk.com/Documentation/Splunk/7.3.0/Indexer/MigratetoSmartStore"&gt;https://docs.splunk.com/Documentation/Splunk/7.3.0/Indexer/MigratetoSmartStore&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;Also, at the time that the index was migrated to SmartStore, any buckets that were in the coldPath continue to  remain in the coldPath.  See &lt;A href="https://docs.splunk.com/Documentation/Splunk/7.3.0/Indexer/SmartStoreindexing"&gt;https://docs.splunk.com/Documentation/Splunk/7.3.0/Indexer/SmartStoreindexing&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2019 19:16:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402773#M3497</guid>
      <dc:creator>Steve_G_</dc:creator>
      <dc:date>2019-06-11T19:16:55Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402774#M3498</link>
      <description>&lt;P&gt;I can confirm that once migration takes place, buckets are &lt;EM&gt;no longer stored on cold&lt;/EM&gt;.&lt;BR /&gt;
EDIT: I tried to force the CM to populate the cache with "cold buckets" but have failed to replicate this behavior. (Ran a search over a small window from months ago on a known index that would have been cold at the time of migration. No colddb population.)&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2019 20:09:57 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402774#M3498</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2019-06-11T20:09:57Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402775#M3499</link>
      <description>&lt;P&gt;Strictly speaking, it's true that the bucket contents will no longer be under coldPath, post-migration, as they are now stored remotely.  But the bucket metadata should still be under coldPath, and bucket contents will get moved to coldPath if required to fulfill a search.&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2019 20:25:17 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402775#M3499</guid>
      <dc:creator>Steve_G_</dc:creator>
      <dc:date>2019-06-11T20:25:17Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402776#M3500</link>
      <description>&lt;P&gt;The entirety of cold storage has 0 files in it. Remains true after running historical searches that would surely end up with some cold data in play.&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2019 20:28:03 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402776#M3500</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2019-06-11T20:28:03Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402777#M3501</link>
      <description>&lt;P&gt;Weird.  Per developer, it's not supposed to work that way.  I'll follow up and report back.&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2019 20:35:12 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402777#M3501</guid>
      <dc:creator>Steve_G_</dc:creator>
      <dc:date>2019-06-11T20:35:12Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402778#M3502</link>
      <description>&lt;P&gt;@SloshBurch - We need a best practice wizard in here.&lt;/P&gt;</description>
      <pubDate>Wed, 12 Jun 2019 19:33:08 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402778#M3502</guid>
      <dc:creator>woodcock</dc:creator>
      <dc:date>2019-06-12T19:33:08Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402779#M3503</link>
      <description>&lt;P&gt;Thanks @woodcock. I hope to tackle smartstore soon and will revisit this at this time.&lt;/P&gt;</description>
      <pubDate>Tue, 16 Jul 2019 14:26:14 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402779#M3503</guid>
      <dc:creator>sloshburch</dc:creator>
      <dc:date>2019-07-16T14:26:14Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402780#M3504</link>
      <description>&lt;P&gt;After migration to SmartStore, the data on coldPath is not automatically removed unless it is forced out through eviction or through the natural aging process.  As Steve pointed out, the coldPath will have metadata stubs and any searches that spans across the cold data will download the data from S3 back to the coldPath.&lt;/P&gt;

&lt;P&gt;Alternatively, after migration, the coldPath location can be changed to some other location (or even homePath) as the idea for migration is to get only a single copy on to S3 and reclaim the space from the warm and cold tiers. &lt;/P&gt;</description>
      <pubDate>Wed, 16 Oct 2019 16:14:30 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402780#M3504</guid>
      <dc:creator>srajarat2</dc:creator>
      <dc:date>2019-10-16T16:14:30Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402781#M3505</link>
      <description>&lt;P&gt;This is spot on, and a behavior I hadn't understood until very recently. Reassigning coldPath to homePath is an excellent idea.&lt;/P&gt;</description>
      <pubDate>Wed, 16 Oct 2019 16:21:39 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402781#M3505</guid>
      <dc:creator>davidpaper</dc:creator>
      <dc:date>2019-10-16T16:21:39Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402782#M3506</link>
      <description>&lt;P&gt;In my logs I see "deletes" files being downloaded, what is the deletes file in the bucket used for? Thanks &lt;/P&gt;</description>
      <pubDate>Mon, 21 Oct 2019 14:03:42 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402782#M3506</guid>
      <dc:creator>jamie00171</dc:creator>
      <dc:date>2019-10-21T14:03:42Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402783#M3507</link>
      <description>&lt;P&gt;That file is where the info is stored to block events from showing up in search that have had "|delete" run against them in the past. &lt;/P&gt;</description>
      <pubDate>Mon, 21 Oct 2019 17:57:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402783#M3507</guid>
      <dc:creator>davidpaper</dc:creator>
      <dc:date>2019-10-21T17:57:55Z</dc:date>
    </item>
    <item>
      <title>Re: SmartStore Behaviors</title>
      <link>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402784#M3508</link>
      <description>&lt;P&gt;Hi David, this is a great session. &lt;BR /&gt;
Today, one Splunk instance identified some issues with smartstore on top of on-prem object storage. It worked normal since smartstore was enabled several months ago. Most of the time, the indexing rate per indexer is about 8-10MB/s. But, while there was a spike (not sure how much yet), indexer processor was stuck and consuming 100% CPU on indexer. All pipelines were blocked and couldn't be recovered. Indexing rate dropped to 2MB/s. They restarted the indexer. It went back to normal with index rate of 16MB/s. &lt;BR /&gt;
Around 20min before the congestion, Some errors like "DatabaseDirectoryManager - failed to open bucket/waif for bucket to be local through CacheManager" started to be reported by indexer. &lt;BR /&gt;
Their hot buckets are on SSD without RAID. &lt;/P&gt;

&lt;P&gt;Any thought on this case? &lt;/P&gt;</description>
      <pubDate>Tue, 12 Nov 2019 06:15:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Knowledge-Management/SmartStore-Behaviors/m-p/402784#M3508</guid>
      <dc:creator>ypeng_splunk</dc:creator>
      <dc:date>2019-11-12T06:15:55Z</dc:date>
    </item>
  </channel>
</rss>

