Deployment Architecture

Getting error on search head regarding replication

vrmandadi
Builder

alt textSearch peer stmsplidxc001.xxx.com has the following message: Indexer Clustering: Too many bucket replication errors to target peer=10.22.12.xxx:8080. Will stop streaming data from hot buckets to this target while errors persist. Check for network connectivity from the cluster peer reporting this issue to the replication port of target peer. If this condition persists, you can temporarily put that peer in manual detention

0 Karma

akocak
Contributor

if indexer or indexer cluster is not down, this message generally indicates high queue utilization in Splunk.
if you have DMC set up:
Settings > Monitoring Console > Indexing > Performance > < 2 Options>
first, if you have a cluster, check deployment first to see what is busy.
then check the instance that is indexing a lot, or high in Splunk queues.
Panel > Median Fill Ratio of Data Processing Queues
Change dropdowns to all queuest and maximum.

if you don't have DMC set up, this search would show you all of your queues. Find your indexer from host name:

index=_internal group=queue 
| eval percfull=round(((current_size_kb/max_size_kb)*100),2) 
| search percfull>80 
| dedup host, name 
| eventstats dc(host) as hostCount 
| table _time host name hostCount current_size_kb max_size_kb percfull 
| rename _time as InfoTime host as Host name as QueueName current_size_kb as CurrentSize max_size_kb as MaxSize percfull as Perc% 
| sort - Perc% 
| convert ctime(InfoTime)
0 Karma

vrmandadi
Builder

InfoTime Host QueueName hostCount CurrentSize MaxSize Perc%
07/26/2018 12:18:07.228 hesplfwd002.xxx.com execprocessorinternalq 14 4095 500 819.00

I got this results for one of the host,is this the issue?

0 Karma

akocak
Contributor

execprocessor queue is %819 percent, is hesplfwd002 an indexer ? execprocessor would not typically stop indexing. you can restart that instance anyways.

for file operation errors in the indexer cluster( typically from antivirus or account permissions) , I can suggest you this search also:

index=_internal (log_level=ERROR OR log_level=WARN) (host=myindexers) (component=PeriodicReapingTimeout OR component=BucketMover OR component=BucketReplicator OR component=CMRepJob OR component=DatabaseDirectoryManager OR component=DispatchCommandProcessor OR component=HttpClientRequest OR component=HttpListener OR component=IniFile OR component=ProcessRunner OR component=TPool OR component=TcpInputProc OR component=TcpOutputFd OR component=IndexerService OR component=S2SFileReceiver) 
| stats values(host) dc(host) last(_raw) count by punct,component 
| fields - punct 

This will get you last error of each type , and components to check for issues. if this points to anything, you could adjust logging levels on that component to see what is failing.

if it is an indexer in cluster, do in CLI to restart :

splunk offline

this will take some time, then:

splunk start

Finally, check bucket status on the Cluster Master. Check which server has copies failing.

0 Karma

vrmandadi
Builder

This hesplfwd002 is actually a heavy forwarder.I ran your search taht you gave and got the below details attached in the screen shot.The only thing I changed in the search is the host=*. The problem is with the indexer hesplidxc001.xxx.com ,how to get this up

0 Karma

akocak
Contributor

Did you restart that instance yet ? I don't see screenshot here

0 Karma

vrmandadi
Builder

Please see in the question I have attached the screen shot there.Yes I have restarted

0 Karma

akocak
Contributor

I can see in 3rd row, you have that indexer is running out of disk space.

0 Karma

renjith_nair
Legend

@vrmandadi,

As mentioned in the error, please check if the connectivity between your peers are not blocked by firewall and splunk service on 10.22.12.xxx is up and running and the port 8080 is not blocked.

---
What goes around comes around. If it helps, hit it with Karma 🙂
0 Karma

vrmandadi
Builder

I checked the peer 10.22.12.xxx:8080 in cluster master which has status as pending and searchable as NO.I tried to restart the peer but still it shows the same thing.I checked the disk usage and found that the /apps is using 100%.Is that the issue ,if so how can we solve this

0 Karma
Get Updates on the Splunk Community!

Earn a $35 Gift Card for Answering our Splunk Admins & App Developer Survey

Survey for Splunk Admins and App Developers is open now! | Earn a $35 gift card!      Hello there,  Splunk ...

Continuing Innovation & New Integrations Unlock Full Stack Observability For Your ...

You’ve probably heard the latest about AppDynamics joining the Splunk Observability portfolio, deepening our ...

Monitoring Amazon Elastic Kubernetes Service (EKS)

As we’ve seen, integrating Kubernetes environments with Splunk Observability Cloud is a quick and easy way to ...