<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Cluster bucket rebalancing: How long is too long? in Deployment Architecture</title>
    <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301979#M11395</link>
    <description>&lt;P&gt;My first few attempts at rebalancing were pretty great. No muss, no fuss. They ran for about 12 hours and like magic my cluster was firing on all cylinders. Beautiful.&lt;/P&gt;

&lt;P&gt;'Stuff happens' and I'm now in the situation again where I've introduced new servers to the the cluster (replacing old). Now I'm way out of balance. "No problem," says I. "Data rebalancing is awesome."&lt;/P&gt;

&lt;P&gt;Not so fast. Literally. I fired it off late Friday night. By Monday morning the process was reading 0.14% done, .01% more than right after starting the process 56 hours previously. By my math that's about 650 days to complete.&lt;/P&gt;

&lt;P&gt;I stopped the process, and restarted it for 1 index only -- 648 buckets using 1 TB of disk. After running for the last 18 hours, it's at 3% complete. So slow as to not really be usable.&lt;/P&gt;

&lt;P&gt;12 servers in the cluster, 4 are new; all are:&lt;/P&gt;

&lt;UL&gt;
&lt;LI&gt;20 core Xeons (2 threads each)&lt;/LI&gt;
&lt;LI&gt;22 1.6TB SSDs&lt;/LI&gt;
&lt;LI&gt;128 GB of RAM&lt;/LI&gt;
&lt;LI&gt; Splunk 6.5.1&lt;/LI&gt;
&lt;LI&gt;Avg load over the weekend was under 2&lt;/LI&gt;
&lt;/UL&gt;

&lt;P&gt;Any suggestions appreciated.&lt;/P&gt;

&lt;P&gt;EDIT:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;&amp;gt; splunk btool server list clustering | grep max_peer
max_peer_build_load = 2
max_peer_rep_load = 5
max_peer_sum_rep_load = 5
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;Jon&lt;/P&gt;</description>
    <pubDate>Tue, 04 Apr 2017 14:06:11 GMT</pubDate>
    <dc:creator>twinspop</dc:creator>
    <dc:date>2017-04-04T14:06:11Z</dc:date>
    <item>
      <title>Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301979#M11395</link>
      <description>&lt;P&gt;My first few attempts at rebalancing were pretty great. No muss, no fuss. They ran for about 12 hours and like magic my cluster was firing on all cylinders. Beautiful.&lt;/P&gt;

&lt;P&gt;'Stuff happens' and I'm now in the situation again where I've introduced new servers to the the cluster (replacing old). Now I'm way out of balance. "No problem," says I. "Data rebalancing is awesome."&lt;/P&gt;

&lt;P&gt;Not so fast. Literally. I fired it off late Friday night. By Monday morning the process was reading 0.14% done, .01% more than right after starting the process 56 hours previously. By my math that's about 650 days to complete.&lt;/P&gt;

&lt;P&gt;I stopped the process, and restarted it for 1 index only -- 648 buckets using 1 TB of disk. After running for the last 18 hours, it's at 3% complete. So slow as to not really be usable.&lt;/P&gt;

&lt;P&gt;12 servers in the cluster, 4 are new; all are:&lt;/P&gt;

&lt;UL&gt;
&lt;LI&gt;20 core Xeons (2 threads each)&lt;/LI&gt;
&lt;LI&gt;22 1.6TB SSDs&lt;/LI&gt;
&lt;LI&gt;128 GB of RAM&lt;/LI&gt;
&lt;LI&gt; Splunk 6.5.1&lt;/LI&gt;
&lt;LI&gt;Avg load over the weekend was under 2&lt;/LI&gt;
&lt;/UL&gt;

&lt;P&gt;Any suggestions appreciated.&lt;/P&gt;

&lt;P&gt;EDIT:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;&amp;gt; splunk btool server list clustering | grep max_peer
max_peer_build_load = 2
max_peer_rep_load = 5
max_peer_sum_rep_load = 5
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;Jon&lt;/P&gt;</description>
      <pubDate>Tue, 04 Apr 2017 14:06:11 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301979#M11395</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2017-04-04T14:06:11Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301980#M11396</link>
      <description>&lt;P&gt;Try to use the lower rebalance_threshold on the master to see if the rebalance performance improves. You can rebalance in multiple waves by increasing the rebalance_threshold.&lt;/P&gt;

&lt;P&gt;&lt;A href="http://docs.splunk.com/Documentation/Splunk/6.5.2/Indexer/Rebalancethecluster#Configure_the_data_rebalancing_threshold" target="_blank"&gt;http://docs.splunk.com/Documentation/Splunk/6.5.2/Indexer/Rebalancethecluster#Configure_the_data_rebalancing_threshold&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 29 Sep 2020 13:32:00 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301980#M11396</guid>
      <dc:creator>somesoni2</dc:creator>
      <dc:date>2020-09-29T13:32:00Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301981#M11397</link>
      <description>&lt;P&gt;Attempting at 0.5. I'll get back with results.&lt;/P&gt;</description>
      <pubDate>Tue, 04 Apr 2017 16:13:43 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301981#M11397</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2017-04-04T16:13:43Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301982#M11398</link>
      <description>&lt;P&gt;What is your retention policy and your replication factor?&lt;/P&gt;</description>
      <pubDate>Thu, 06 Apr 2017 15:46:18 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301982#M11398</guid>
      <dc:creator>mattlucas719</dc:creator>
      <dc:date>2017-04-06T15:46:18Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301983#M11399</link>
      <description>&lt;P&gt;Hah! @twinspop I seem to be following you from &lt;A href="https://answers.splunk.com/answers/476015/whats-the-best-method-to-updatereplace-indexer-clu.html"&gt;https://answers.splunk.com/answers/476015/whats-the-best-method-to-updatereplace-indexer-clu.html&lt;/A&gt; &lt;/P&gt;

&lt;P&gt;I am in the same spot now and indexer rebalance is painfully slow. Did you find out the cause for this? &lt;/P&gt;</description>
      <pubDate>Mon, 26 Jun 2017 23:22:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301983#M11399</guid>
      <dc:creator>sk314</dc:creator>
      <dc:date>2017-06-26T23:22:55Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301984#M11400</link>
      <description>&lt;P&gt;My cold storage was on NAS. Since having all stages (hot, warm, cold) on local drives, rebalance is plenty fast.&lt;/P&gt;</description>
      <pubDate>Tue, 27 Jun 2017 01:35:32 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301984#M11400</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2017-06-27T01:35:32Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301985#M11401</link>
      <description>&lt;P&gt;The problem (apparently) was related to COLD storage being on NAS. Since restructuring our storage plans to get everything on local drives, data rebalance is a quick process.&lt;/P&gt;</description>
      <pubDate>Tue, 27 Jun 2017 01:36:28 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301985#M11401</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2017-06-27T01:36:28Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301986#M11402</link>
      <description>&lt;P&gt;Hmmm - we have all our storage on local drives - except frozen/archive - which is on NFS. Does frozen/archive drive count? &lt;/P&gt;</description>
      <pubDate>Tue, 27 Jun 2017 01:39:36 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301986#M11402</guid>
      <dc:creator>sk314</dc:creator>
      <dc:date>2017-06-27T01:39:36Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster bucket rebalancing: How long is too long?</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301987#M11403</link>
      <description>&lt;P&gt;No, frozen data isn't involved in data rebalance.&lt;/P&gt;</description>
      <pubDate>Tue, 27 Jun 2017 01:42:54 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Cluster-bucket-rebalancing-How-long-is-too-long/m-p/301987#M11403</guid>
      <dc:creator>twinspop</dc:creator>
      <dc:date>2017-06-27T01:42:54Z</dc:date>
    </item>
  </channel>
</rss>

