I am trying to ingest some json data into a new Splunk Cloud instance, with a custom sourcetype, but I keep getting duplicate data in the search results. This seems to be an extremely common problem, based on the number of old posts, but none of them seem to address the Cloud version.
I have a JSON file that looks like this:
{ "RowNumber": 1, "ApplicationName": "177525278", "ClientProcessID": 114889, "DatabaseName": "1539703986", "StartTime": "2024-07-30 12:15:13" } |
When I add a .json file to the folder, it is ingested and the events show up in the cloud instance, under the right correct centraladmin_errormaster index, and with the sourcetype=errormaster.
However, the fields all have duplicate values.
If it switch it to the built-in _json type it works fine. I have some field extractions I want to add, which is why I wanted a custom type.
I'm guessing this is something obvious to the Cloud experts, but I am an accidental Splunk Admin with very little experience, so any help you can offer would be appreciated.
Our partner had me delete the props.conf file on the Universal Forwarder and that seems to have solved the issue. I still don't understand why the _json type works correctly while a copy of the _json type does not, but for now this issue is resolved.
This behaviour is typically a result of having both INDEXED_EXTRACTIONS as well as search-time extractions active. Can't say what your effective config is so don't know for sure why it's happening.
Our partner had me delete the props.conf file on the Universal Forwarder and that seems to have solved the issue. I still don't understand why the _json type works correctly while a copy of the _json type does not, but for now this issue is resolved.