Getting Data In

Is there a way to monitor the tcp write to Indexers and check the amount of data an Indexer receives on an average?

Harishma
Communicator

We have an environment where we directly write data to Splunk indexers via TCP inputs.
The reason for this kind of set up is kafka consumers consume data from Kafka and later this is writtent into splunk.

Sometimes this tcp takes too much time and data doesnt get into Indexers.

Is there a way to monitor if this data really gets into splunk?

Also how can we find out how much data is writtent into any indexer at any given time? ( we do not have a clustered Indexer and would like to know the how much gets into every Indexer - we have like 100 Indexers)

0 Karma
1 Solution

VatsalJagani
SplunkTrust
SplunkTrust

@Harishma,

Sometimes this tcp takes too much time and data doesn't get into Indexers.
I think you must be using TCP input in Splunk on some port. In that case, all data received on the port will surely be indexed by the Splunk. Make sure on the sendor machine or on receiver there is no throughput set for network data transfer at the operating system level.

Is there a way to monitor if this data really gets into Splunk?
Use query index=<your data index> | stats count to check all the events are received or not. This query gives the number of events.

If you have multiple indexer in cluster then use query index=<your data index> | stats count by splunk_server to check events count on different indexers. This represents data distribution in splunk clustered enviorment.

Hope this helps!!!

View solution in original post

ddrillic
Ultra Champion

Interesting, we had yesterday a similar question about syslog at How to calculate volume of syslog traffic on syslog-ng server

0 Karma

martynoconnor
Communicator

You can use the following search on the licence master to see how much each indexer has indexed:

index=_internal sourcetype=splunkd source=*license_usage.log* 
| fields b idx splunk_server
| eval MB=b/1024/1024
| stats sum(MB) as bytes by idx splunk_server
0 Karma

VatsalJagani
SplunkTrust
SplunkTrust

@Harishma,

Sometimes this tcp takes too much time and data doesn't get into Indexers.
I think you must be using TCP input in Splunk on some port. In that case, all data received on the port will surely be indexed by the Splunk. Make sure on the sendor machine or on receiver there is no throughput set for network data transfer at the operating system level.

Is there a way to monitor if this data really gets into Splunk?
Use query index=<your data index> | stats count to check all the events are received or not. This query gives the number of events.

If you have multiple indexer in cluster then use query index=<your data index> | stats count by splunk_server to check events count on different indexers. This represents data distribution in splunk clustered enviorment.

Hope this helps!!!

Harishma
Communicator

HI @VatsalJagani,

Sorry for late reply
Yes you're right, tcp takes too much to write, I'll check or the throughput part may be. Never looked into it.
But are you aware of why TCP collapses occur?

0 Karma

VatsalJagani
SplunkTrust
SplunkTrust

It may have many reasons, like network issues, throughput, etc.

0 Karma
Get Updates on the Splunk Community!

Earn a $35 Gift Card for Answering our Splunk Admins & App Developer Survey

Survey for Splunk Admins and App Developers is open now! | Earn a $35 gift card!      Hello there,  Splunk ...

Continuing Innovation & New Integrations Unlock Full Stack Observability For Your ...

You’ve probably heard the latest about AppDynamics joining the Splunk Observability portfolio, deepening our ...

Monitoring Amazon Elastic Kubernetes Service (EKS)

As we’ve seen, integrating Kubernetes environments with Splunk Observability Cloud is a quick and easy way to ...