Getting Data In

Ingestion via Spark



We are using Spark Apps to ingest the data into Splunk. For that, we are referring to documentation. 

We are planning to use attachWith. Is there any other/better way to achieve the same.

There are a lot of examples in Java. Is there any better documentation/example to ingest bulk data(In TBs) via Spark Apps. 

Labels (1)
Tags (2)
0 Karma
Get Updates on the Splunk Community!

Observability | How to Think About Instrumentation Overhead (White Paper)

Novice observability practitioners are often overly obsessed with performance. They might approach ...

Cloud Platform | Get Resiliency in the Cloud Event (Register Now!)

IDC Report: Enterprises Gain Higher Efficiency and Resiliency With Migration to Cloud  Today many enterprises ...

The Great Resilience Quest: 10th Leaderboard Update

The tenth leaderboard update (11.23-12.05) for The Great Resilience Quest is out >> As our brave ...