Hi guys.
Why Splunk have many errors in log file and what can I do in this situation?
05-17-2019 18:58:08.036 +0300 WARN DistributedPeerManager - Cannot determine a latest common bundle, search may be blocked
05-17-2019 18:58:08.036 +0300 WARN DistributedPeerManager - Cannot determine a latest common bundle, search may be blocked
05-17-2019 18:58:08.037 +0300 WARN DistributedPeerManager - Cannot determine a latest common bundle, search may be blocked
I've encountered this before, especially on new SHC builds.
You'll need to perform a manual / destructive resync in order to get them properly clustered again.
splunk resync shcluster-replicated-config
If that does not work you will likely need to bootstrap a new captain, then add (init) the other members back in.
splunk bootstrap shcluster-captain -servers_list "<URI>:<management_port>,<URI>:<management_port>,..." -auth <username>:<password>
It is possible that there are other things going on that is causing this error than what is stated above. Since I identified a unique root cause I wanted to share with all. The last bullet below was what worked for me but the below bullets represents a summary of recommended steps to get to root cause for this.
If this error is been generated on the cluster master. Go to Settings > Distributed Peers and verify the health of the indexers, it's possible that the remote credentials have expired or has changed. Click on each of the peer nodes and re-authenticate. This should fix the issue.
/opt/splunk/bin/splunk bootstrap shcluster-captain -servers_list "https://splunk-sh11:8089,https://splunk-sh21:8089" -auth admin:XXXXXX
server=https://splunk-sh11:8089, error=This node seems to have already joined another cluster with below members: 'https://splunk-sh11:8089,https://splunk-sh21:8089'.
First remove the member from the old cluster. Then run 'splunk clean raft' on the member to reuse it in a new cluster; server=https://splunk-sh21:8089, error=This node seems to have already joined another cluster with below members: 'https://splunk-sh11:8089,https://splunk-sh21:8089'.
First remove the member from the old cluster. Then run 'splunk clean raft' on the member to reuse it in a new cluster;
Hello again.
If I already have a configured cluster, will the initial bootstrap of the cluster master result in the loss of all settings and users?
I've encountered this before, especially on new SHC builds.
You'll need to perform a manual / destructive resync in order to get them properly clustered again.
splunk resync shcluster-replicated-config
If that does not work you will likely need to bootstrap a new captain, then add (init) the other members back in.
splunk bootstrap shcluster-captain -servers_list "<URI>:<management_port>,<URI>:<management_port>,..." -auth <username>:<password>
[root@splunk-sh21 certs]# /opt/splunk/bin/splunk resync shcluster-replicated-config
The member has been synced to the latest replicated configurations on the captain.
But I've same troubles
05-20-2019 10:57:18.405 +0300 WARN DistributedPeerManager - Cannot determine a latest common bundle, search may be blocked
I think your best bet is to rebuild the SHC altogether. Remove all members from the cluster and cycle Splunk. Then bootstrap one of the nodes as captain with the command I posted previously.
can you login to the DS, push the latest bundle and do splunk apply cluster-bundle from the Cluster master to all your peers. Do paste the errors which you get post doing this..
Always try to check the CM bundle details and compare if the latest active bundle in the peers is same as the CM