Deployment Architecture

How do I make the cluster manager rendundant ?

Ulfb
Explorer

We had an issue with the cluster manager, that now resolved.

How do one design away the cluster manager as an single point of failure ?

Tags (3)

theunf
Communicator

Update: using a new installed Splunk instance made the difference instead of setting it as a MAster thru clustering gui :

  • new 6.1.3 x64 install;
  • setting initial password, changing licensing mode, set https and port;
  • stop new instance;
  • making changes on server.conf;
  • untar .tgz file with $SPLUNK_HOME/etc/master-apps/ of the old master;
  • start new master;
  • stop all Indexers;
  • change master_uri on [clustering] stanza of server.conf of all Indexers;
  • start each Indexer at a time;

I guess how to create a full automated solution using linux´s Heartbeat and a bunch of scripts, for example ;(

0 Karma

theunf
Communicator

Hi all,

I´m testing this master-stanby on a lab and having problems with the indexers startup and error messages (at the end).

The article that i´m following is on the docs : http://docs.splunk.com/Documentation/Splunk/6.1.3/Indexer/Handlemasternodefailure

Basicly, I made a new server, change it to Master role than stopped the process. Copied master-apps thru rsync , then tested again using tar on the old master and untarring on the new ... same errors.

For the record... with the Search Heads i only needed to point to the new URL thru the Settings->Clustering config.

10-02-2014 15:24:26.434 -0300 INFO CMBundleMgr - Downloaded bundle path=/opt/splunk/var/run/splunk/cluster/remote-bundle/4c74b4f7cd208b4fd98ca7698c1a77db1412274266.bundle time_taken_ms=9.
10-02-2014 15:24:26.434 -0300 INFO CMBundleMgr - untarring bundle=/opt/splunk/var/run/splunk/cluster/remote-bundle/4c74b4f7cd208b4fd98ca7698c1a77db-1412274266.bundle
10-02-2014 15:24:26.441 -0300 INFO ClusterBundleValidator - Validating bundle path=/opt/splunk/var/run/splunk/cluster/remote-bundle/4c74b4f7cd208b4fd98ca7698c1a77db-1412274266/apps
10-02-2014 15:24:26.460 -0300 INFO CMBundleMgr - Removed the untarred bundle folder=/opt/splunk/var/run/splunk/cluster/remote-bundle/4c74b4f7cd208b4fd98ca7698c1a77db-1412274266
10-02-2014 15:24:26.460 -0300 INFO CMBundleMgr - Removed the bundle downloaded from master to '/opt/splunk/var/run/splunk/cluster/remote-bundle/4c74b4f7cd208b4fd98ca7698c1a77db-1412274266.bundle'

0 Karma

Drainy
Champion

I think the answer here is that Splunk needs to enhance the HA features of the cluster master, have a look at my older question here;
http://answers.splunk.com/answers/65397/splunk-v5-clustering-and-ha

There is some good detail in the answers there from other people.

0 Karma

mahamed_splunk
Splunk Employee
Splunk Employee

Agreed. Cluster Master HA is on our roadmap.

0 Karma

BigRoundOne
New Member

We experience the same time to recover a Cluster Master as Dolxor, about 3-4 hours for 8 indexers of 800GB each. It takes about 30-45 minutes per cluster peer to finish its processing and the Cluster master processes each sequentially.

0 Karma

dolxor
Path Finder

In this incident we had 4 indexers with about 3-4 TB of data each.

0 Karma

mahamed_splunk
Splunk Employee
Splunk Employee

Cluster master keeps only the meta data, so in order to bring up secondary master we need some basic static files like server.conf. All other information such as indexes, buckets etc can be dynamically rebuild by secondary master. More on this is documented here
http://docs.splunk.com/Documentation/Splunk/5.0/Indexer/Configurethemaster#Configure_a_stand-by_mast...

Dolxor, I'm surprised that it took 4 - 6 hours for the cluster master to catch up. If all the peers are available and you just did a cluster master restart, it should get back to normal state in few mins. If you notice this again, please open a support case so we can get to the bottom of the issue

dolxor
Path Finder

Splunk Support was involved in the case. They assisted us by phone/web-conference.

0 Karma

dolxor
Path Finder

What Splunk should allow was for us to use the same license, index the data to one local and one off-site Splunk Cluster with same hardware. If your local Cluster Master dies horribly, your search head could use the off-site Cluster Master towards the off-site Cluster to allow for continued operations while you fix your local cluster master.

You can do that today, it is possible. But the cost is double. For some of us, cost is a reality we cannot ignore.

0 Karma

omniit
Explorer

Clustering is always just 1 step in reducing your exposure to failures. A cluster will in practice only protect you from technical failures with hardware. It does nothing to prevent logical corruption or other logical issues within the application or its logic. If you would like to have full disaster tolerancy (physical and logical) for you entire Splunk setup, good luck... I doubt it will be possible at all. Maybe when using various techniques combined with true non-stop hardware (HP Non-Stop / Himalaya, type of equipment) clusters, then maybe it is possible but the risk vs benefits vs cost, calculation will be a hard sell within any organisation. Perhaps Splunk US has some reference architecture documentation about the HA solution that Apple is using.

0 Karma

dolxor
Path Finder

If your (primary) Cluster Master dies horribly, would all this fix this? The new hot-standby (or whatever) Cluster Master then needs to take over the role. Last time this happened to us (the primary Cluster Master was fixed and booted up again) - it took us about 4-6 hours of nervous waiting time for the cluster master to 'fix' stuff in the cluster before everything was neat and nice again.

A hot-standby cluster master would still need to do all this "fixing" stuff before you can use the cluster for searching (indexing was still OK during our downtime). So you do not fix the underlying problem, and the cluster master is still a very serious single-point-of-failure in a Splunk cluster scenario.

0 Karma

omniit
Explorer

If you are looking for a high available solution, without OS clustering there is also the option of running a VM. If your virtualization platform is already highly available and you have to option of using it, why not host the cluster manager role on there? Running a hot-standby VM is possible with either VMware ESX, Hyper-V, XenServer and I would assume most other solutions also... and don't forget, Splunk failover functionality is impacted by the loss of the cluster manager, but no 'real' functionality is lost with a momentary loss of the CM. So a hot or warm stand-by CM would probably be enough to meet most HA demands.

0 Karma

omniit
Explorer

The cluster manager is only handling some basic management tasks, so no real heavy loads. Since combined setups are supported by Splunk, it would not be a problem to only host you Cluster Manager on FO-Cluster and run all 'real' loads on a different OS. Since I am a former MS employee, Windows clustering is within my comfort zone and thus I suggested it. I would however in all fairness have to say, that if you don't have in-depth Windows knowledge and you are already used to working with some other OS, it would of course make more sense to just run a cluster solution based on you own preferred OS. I am however assuming that the concept of a FO-Cluster in Windows, has a equivalent solution (functionality-wise) in the OS you are intending to use.

Regards,

Martijn (Omni-it is just my company name. )

0 Karma

Ulfb
Explorer

I have 10 year background with windows cluster, and have mixed feelings.

Both in the preformance we got from windows as base os for Splunk, and the complexisty clustering i windows gives.

Second, we was hit by an sw issue that created an race like issue in splunk.

Im looking for good designs/ideas that might address this.

We would be equaly stuck with an os clustering.

Thanks for all input so far.

0 Karma

omniit
Explorer

You could setup a Windows failover cluster specific for this role and define the Splunk process as a custom service. It will then failover when your primairy node or a dependency (Network/Disk) fails.

0 Karma

Ulfb
Explorer

Anyone realy running it on windows bokses ? Wow..

Well, thanks for pointing out that it is possible.

Do windows cluster handle filelocking and would handle an bug in cluster manager ?

Windows cluistering have some many possible options to fail, and add way to much complexity.

I did consider something like beowulf cluster, but dropped it.

Is there no arcitectural way to design ourself away ?

0 Karma
Get Updates on the Splunk Community!

New This Month in Splunk Observability Cloud - Metrics Usage Analytics, Enhanced K8s ...

The latest enhancements across the Splunk Observability portfolio deliver greater flexibility, better data and ...

Alerting Best Practices: How to Create Good Detectors

At their best, detectors and the alerts they trigger notify teams when applications aren’t performing as ...

Discover Powerful New Features in Splunk Cloud Platform: Enhanced Analytics, ...

Hey Splunky people! We are excited to share the latest updates in Splunk Cloud Platform 9.3.2408. In this ...