Getting Data In

Splunk Indexer Parsing Queue Blocking

ayeheyis
Explorer

Hello,

I just recently restarted my splunk enterprise instance in order to add an app and once it was back up, i noticed that one of the health checks was failing. 

Screen Shot 2022-02-01 at 9.16.38 PM.png

Also no new logs were showing up in the search. 

I looked at the monitoring console and noticed the parsing queue was full.

Screen Shot 2022-02-01 at 9.19.58 PM.png

I also checked the metrics.log and saw some of the queues were full.

Screen Shot 2022-02-01 at 9.36.03 PM.png

If I'm understanding the data pipeline hierarchy correctly, it's the parsing queue that's actually blocked and causing the other queues to be blocked. 

I also checked the splunkd.log and didn't really anything that seemed related. There were some SSL errors which didn't seem related. And this other error:

 

ERROR HttpInputDataHandler - Failed processing http input, token name=kube, channel=n/a, source_IP=172.17.8.66, reply=9, events_processed=4, http_input_body_size=7256, parsing_err="Server is busy"

 

but that seems to be a result of the full queue. 

I looked into my resource usage from the monitoring console and top tool and neither cpu or mem go higher than 50% utilization.

I also restarted splunk multiple times but the queue always goes to 100% instantly. I did notice a warning on startup:

Bad regex value: '(::)?...', of param: props.conf / [(::)?...]; why: this regex is likely to apply to all data and may break summary indexing, among other Splunk features.

However, I didn't make any changes to props.conf and everything was working before I restarted the first time so I assume this is not related.

Not sure what else to try. Any help would be greatly appreciated!

Labels (1)
0 Karma
1 Solution

isoutamo
SplunkTrust
SplunkTrust

Hi

you said that you just installed a new app there? Probably it contains something which broke the data pipeline? 

You should start with the next documents to understanding what can cause this issue:

As the parsing Queue is full I start to looking those actions which are happening in parsing pipeline

---8<---

Parsing
During the parsing segment, Splunk software examines, analyzes, and transforms the data. This is also known as event processing. It is during this phase that Splunk software breaks the data stream into individual events.The parsing phase has many sub-phases:

  • Breaking the stream of data into individual lines.
  • Identifying, parsing, and setting timestamps.
  • Annotating individual events with metadata copied from the source-wide keys.
  • Transforming event data and metadata according to regex transform rules.

---8<--

There is unofficial trick to use wildcard on props.conf. Just use (?::){0} in beginning of your e.g. sourcetype stanza. Probably someone has remember this wrongly and put those in wrong order?

r. Ismo

View solution in original post

isoutamo
SplunkTrust
SplunkTrust

Hi

you said that you just installed a new app there? Probably it contains something which broke the data pipeline? 

You should start with the next documents to understanding what can cause this issue:

As the parsing Queue is full I start to looking those actions which are happening in parsing pipeline

---8<---

Parsing
During the parsing segment, Splunk software examines, analyzes, and transforms the data. This is also known as event processing. It is during this phase that Splunk software breaks the data stream into individual events.The parsing phase has many sub-phases:

  • Breaking the stream of data into individual lines.
  • Identifying, parsing, and setting timestamps.
  • Annotating individual events with metadata copied from the source-wide keys.
  • Transforming event data and metadata according to regex transform rules.

---8<--

There is unofficial trick to use wildcard on props.conf. Just use (?::){0} in beginning of your e.g. sourcetype stanza. Probably someone has remember this wrongly and put those in wrong order?

r. Ismo

ayeheyis
Explorer

Just to close the loop on this, as I was going thru the logs I noticed the app I recently installed was still enabled even tho I could've sworn I disabled it. I disabled the app and restarted using the web interface. Not sure if that matters but when I did it thru the CLI, everything seems to be working again.

Thanks for all your help!

0 Karma

johnhuang
Motivator

 A bottleneck in the pipeline (e.g. parsing queue) will cause the upstream buffer to be maxed out. You can:

  • Increase the size limit of your queue buffer will allow more time for your CPU to process backlogs.
  • Throw more hardware at it, e.g. offload to another heavy forwarder.
  • Optimize the regex in your parsing queue (props.conf transform.conf), starting with your largest sourcetypes.

 Here's a query to calculate the average, peak, and max size limit for the different queues. You can use this to determine which queue needs to be increased.

 

index=_internal source=*metrics.log group=queue host=<splunk_host> max_size_kb=* earliest=-1d
| stats max(max_size_kb) AS max_limit_kb max(current_size_kb) AS peak_size_kb avg(current_size_kb) AS avg_size_kb by name
| eval avg_size_kb=ROUND(avg_size_kb, 2)

 

 

ayeheyis
Explorer

Thanks for the response. Do you know which properties I should set? I went thru the limits.conf documentation and I couldn't figure which properties relate to the queue sizes.

Also I'm not able to see any logs for the index _internal. I'm assuming it's caused by the blockage in the data pipeline. 

0 Karma

isoutamo
SplunkTrust
SplunkTrust
In the first step is to figure out what is the reason for that issue. Then you can start to fix it e.g. with limits or much probably fixing some other configurations like setting timestamp recognition on props.conf.
Get Updates on the Splunk Community!

Earn a $35 Gift Card for Answering our Splunk Admins & App Developer Survey

Survey for Splunk Admins and App Developers is open now! | Earn a $35 gift card!      Hello there,  Splunk ...

Continuing Innovation & New Integrations Unlock Full Stack Observability For Your ...

You’ve probably heard the latest about AppDynamics joining the Splunk Observability portfolio, deepening our ...

Monitoring Amazon Elastic Kubernetes Service (EKS)

As we’ve seen, integrating Kubernetes environments with Splunk Observability Cloud is a quick and easy way to ...